2025-04-02 04:14:02,849 [ 807451 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-02 04:14:02,849 [ 807451 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-02 04:14:02,849 [ 807451 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-02 04:14:02,849 [ 807451 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_l4sdy9 --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=8b2301119731 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss" altinityinfra/integration-tests-runner:2165613c5fcd '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [100 items] scheduling tests via LoadFileScheduling Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3] test_storage_url/test.py::test_file_formats[CSV] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_ssh_keys_authentication/test.py::test_ecdsa test_system_queries/test.py::test_DROP_DNS_CACHE test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_table_function_mongodb_legacy/test.py::test_auth_source[False] test_storage_redis/test.py::test_create_table test_system_merges/test.py::test_merge_simple[] test_system_start_stop_listen/test.py::test_all_protocols Stdout:1 Stdout:1 No running containers No running containers Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Stdout:1 Stdout:1 No running containers No running containers Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Stdout:1 Stdout:1 Stdout:1 Stdout:1 Stdout:1 Stdout:1 Stdout:1 Stdout:1 No running containers No running containers No running containers No running containers No running containers No running containers Pruning Docker networks Pruning Docker networks Pruning Docker networks Pruning Docker networks No running containers No running containers Stdout:1 Command:[docker network prune --force] Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Pruning Docker networks Stdout:1 Command:[docker network prune --force] Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Command:[docker network prune --force] Command:[docker network prune --force] No running containers No running containers Stdout:1 Stdout:1 Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] No running containers Command:[docker network prune --force] No running containers Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Stdout:1 Stdout:1 No running containers No running containers Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Stdout:1 Stdout:1 No running containers No running containers Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_ssh_keys_authentication/test.py Running tests in /ClickHouse/tests/integration/test_ssh_keys_authentication/test.py Cluster start called. is_up=False Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 5adb7eee8142 ENV HOSTNAME 5adb7eee8142 ENV SHLVL 0 ENV SHLVL 0 ENV HOME /root ENV HOME /root Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV OLDPWD / ENV OLDPWD / Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e Stderr:Error response from daemon: a prune operation is already running Stderr:Error response from daemon: a prune operation is already running ENV UBSAN_OPTIONS print_stacktrace=1 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV UBSAN_OPTIONS print_stacktrace=1 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss Exitcode:1 Exitcode:1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Running tests in /ClickHouse/tests/integration/test_system_start_stop_listen/test.py Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Running tests in /ClickHouse/tests/integration/test_system_start_stop_listen/test.py ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse Cluster start called. is_up=False ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Cluster start called. is_up=False ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 Running tests in /ClickHouse/tests/integration/test_storage_redis/test.py Running tests in /ClickHouse/tests/integration/test_storage_redis/test.py ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge Cluster start called. is_up=False Cluster start called. is_up=False ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV LC_CTYPE C.UTF-8 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV INTEGRATION_TESTS_RUN_ID 0 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV WORKER_FREE_PORTS 30150 30151 30152 30153 30154 30155 30156 30157 30158 30159 30160 30161 30162 30163 30164 30165 30166 30167 30168 30169 30170 30171 30172 30173 30174 30175 30176 30177 30178 30179 30180 30181 30182 30183 30184 30185 30186 30187 30188 30189 30190 30191 30192 30193 30194 30195 30196 30197 30198 30199 ENV WORKER_FREE_PORTS 30150 30151 30152 30153 30154 30155 30156 30157 30158 30159 30160 30161 30162 30163 30164 30165 30166 30167 30168 30169 30170 30171 30172 30173 30174 30175 30176 30177 30178 30179 30180 30181 30182 30183 30184 30185 30186 30187 30188 30189 30190 30191 30192 30193 30194 30195 30196 30197 30198 30199 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca Setup Keeper Setup Keeper Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV PYTEST_XDIST_WORKER gw3 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV PYTEST_XDIST_WORKER gw3 Cluster name: project_name:roottests3zerocopyreplication-gw0. Added instance name:node1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env', '--project-name', 'roottests3zerocopyreplication-gw0', '--file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:roottests3zerocopyreplication-gw0. Added instance name:node1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env', '--project-name', 'roottests3zerocopyreplication-gw0', '--file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_XDIST_WORKER_COUNT 10 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV PYTEST_CURRENT_TEST test_table_function_mongodb_legacy/test.py::test_auth_source[False] (setup) ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV PYTEST_CURRENT_TEST test_table_function_mongodb_legacy/test.py::test_auth_source[False] (setup) Cluster name: project_name:roottests3zerocopyreplication-gw0. Added instance name:node2 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env', '--project-name', 'roottests3zerocopyreplication-gw0', '--file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml', '--file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 Cluster name: project_name:roottests3zerocopyreplication-gw0. Added instance name:node2 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env', '--project-name', 'roottests3zerocopyreplication-gw0', '--file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml', '--file', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 5adb7eee8142 ENV HOSTNAME 5adb7eee8142 Starting cluster... Starting cluster... ENV SHLVL 0 ENV SHLVL 0 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV HOME /root ENV HOME /root Running tests in /ClickHouse/tests/integration/test_s3_zero_copy_replication/test.py Running tests in /ClickHouse/tests/integration/test_s3_zero_copy_replication/test.py ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV OLDPWD / ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_HELPER_TAG 5dc43a6382f0 Cluster start called. is_up=False Cluster start called. is_up=False ENV PYTHONUNBUFFERED 1 ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV PYTHONUNBUFFERED 1 ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV UBSAN_OPTIONS print_stacktrace=1 ENV UBSAN_OPTIONS print_stacktrace=1 ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 CLUSTER INIT base_config_dir:/clickhouse-config CLUSTER INIT base_config_dir:/clickhouse-config ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV HOSTNAME 5adb7eee8142 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV COMPOSE_HTTP_TIMEOUT 600 ENV HOSTNAME 5adb7eee8142 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV SHLVL 0 ENV SHLVL 0 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV HOME /root ENV HOME /root ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV OLDPWD / ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV OLDPWD / ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CHANNEL stable ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_CLIENT_TIMEOUT 300 ENV PYTHONUNBUFFERED 1 ENV PYTHONUNBUFFERED 1 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e Cluster name: project_name:roottesttablefunctionmongodblegacy-gw3. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env', '--project-name', 'roottesttablefunctionmongodblegacy-gw3', '--file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV UBSAN_OPTIONS print_stacktrace=1 ENV UBSAN_OPTIONS print_stacktrace=1 Cluster name: project_name:roottesttablefunctionmongodblegacy-gw3. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env', '--project-name', 'roottesttablefunctionmongodblegacy-gw3', '--file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV PWD /ClickHouse/tests/integration ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV COMPOSE_HTTP_TIMEOUT 600 ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config Running tests in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config Running tests in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py ENV TZ Etc/UTC ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV TZ Etc/UTC ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV DOCKER_BASE_TAG 8b2301119731 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_BASE_TAG 8b2301119731 ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV DOCKER_CHANNEL stable ENV DOCKER_CHANNEL stable ENV LC_CTYPE C.UTF-8 ENV LC_CTYPE C.UTF-8 ENV DOCKER_CLIENT_TIMEOUT 300 ENV INTEGRATION_TESTS_RUN_ID 0 Cluster start called. is_up=False ENV INTEGRATION_TESTS_RUN_ID 0 Cluster start called. is_up=False ENV DOCKER_CLIENT_TIMEOUT 300 ENV WORKER_FREE_PORTS 30100 30101 30102 30103 30104 30105 30106 30107 30108 30109 30110 30111 30112 30113 30114 30115 30116 30117 30118 30119 30120 30121 30122 30123 30124 30125 30126 30127 30128 30129 30130 30131 30132 30133 30134 30135 30136 30137 30138 30139 30140 30141 30142 30143 30144 30145 30146 30147 30148 30149 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV WORKER_FREE_PORTS 30100 30101 30102 30103 30104 30105 30106 30107 30108 30109 30110 30111 30112 30113 30114 30115 30116 30117 30118 30119 30120 30121 30122 30123 30124 30125 30126 30127 30128 30129 30130 30131 30132 30133 30134 30135 30136 30137 30138 30139 30140 30141 30142 30143 30144 30145 30146 30147 30148 30149 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PYTEST_XDIST_WORKER gw2 ENV PYTEST_XDIST_WORKER gw2 ENV PWD /ClickHouse/tests/integration ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PWD /ClickHouse/tests/integration ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection (setup) ENV PYTEST_CURRENT_TEST test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection (setup) ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV TZ Etc/UTC CLUSTER INIT base_config_dir:/clickhouse-config CLUSTER INIT base_config_dir:/clickhouse-config ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV DOCKER_BASE_TAG 8b2301119731 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV DOCKER_BASE_TAG 8b2301119731 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV LC_CTYPE C.UTF-8 Cluster name:cluster project_name:rootteststorageazureblobstoragecluster-gw2. Added instance name:node_0 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env', '--project-name', 'rootteststorageazureblobstoragecluster-gw2', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV INTEGRATION_TESTS_RUN_ID 0 Cluster name:cluster project_name:rootteststorageazureblobstoragecluster-gw2. Added instance name:node_0 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env', '--project-name', 'rootteststorageazureblobstoragecluster-gw2', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV INTEGRATION_TESTS_RUN_ID 0 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV WORKER_FREE_PORTS 30350 30351 30352 30353 30354 30355 30356 30357 30358 30359 30360 30361 30362 30363 30364 30365 30366 30367 30368 30369 30370 30371 30372 30373 30374 30375 30376 30377 30378 30379 30380 30381 30382 30383 30384 30385 30386 30387 30388 30389 30390 30391 30392 30393 30394 30395 30396 30397 30398 30399 ENV WORKER_FREE_PORTS 30350 30351 30352 30353 30354 30355 30356 30357 30358 30359 30360 30361 30362 30363 30364 30365 30366 30367 30368 30369 30370 30371 30372 30373 30374 30375 30376 30377 30378 30379 30380 30381 30382 30383 30384 30385 30386 30387 30388 30389 30390 30391 30392 30393 30394 30395 30396 30397 30398 30399 ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca Cluster name:cluster project_name:rootteststorageazureblobstoragecluster-gw2. Added instance name:node_1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env', '--project-name', 'rootteststorageazureblobstoragecluster-gw2', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name:cluster project_name:rootteststorageazureblobstoragecluster-gw2. Added instance name:node_1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env', '--project-name', 'rootteststorageazureblobstoragecluster-gw2', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV PYTEST_XDIST_WORKER gw7 ENV PYTEST_XDIST_WORKER gw7 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_XDIST_WORKER_COUNT 10 Running tests in /ClickHouse/tests/integration/test_system_merges/test.py ENV PYTEST_CURRENT_TEST test_system_queries/test.py::test_DROP_DNS_CACHE (setup) Running tests in /ClickHouse/tests/integration/test_system_merges/test.py ENV PYTEST_CURRENT_TEST test_system_queries/test.py::test_DROP_DNS_CACHE (setup) Cluster name:cluster project_name:rootteststorageazureblobstoragecluster-gw2. Added instance name:node_2 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env', '--project-name', 'rootteststorageazureblobstoragecluster-gw2', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name:cluster project_name:rootteststorageazureblobstoragecluster-gw2. Added instance name:node_2 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env', '--project-name', 'rootteststorageazureblobstoragecluster-gw2', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster start called. is_up=False Cluster start called. is_up=False Running tests in /ClickHouse/tests/integration/test_storage_azure_blob_storage/test_cluster.py Running tests in /ClickHouse/tests/integration/test_storage_azure_blob_storage/test_cluster.py CLUSTER INIT base_config_dir:/clickhouse-config CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster start called. is_up=False Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Cluster name: project_name:roottestsystemqueries-gw7. Added instance name:ch1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env', '--project-name', 'roottestsystemqueries-gw7', '--file', '/ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:roottestsystemqueries-gw7. Added instance name:ch1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env', '--project-name', 'roottestsystemqueries-gw7', '--file', '/ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Running tests in /ClickHouse/tests/integration/test_system_queries/test.py Running tests in /ClickHouse/tests/integration/test_system_queries/test.py Running tests in /ClickHouse/tests/integration/test_storage_url/test.py Running tests in /ClickHouse/tests/integration/test_storage_url/test.py Cluster start called. is_up=False Cluster start called. is_up=False Cluster start called. is_up=False Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 5adb7eee8142 ENV HOSTNAME 5adb7eee8142 ENV SHLVL 0 ENV SHLVL 0 ENV HOME /root ENV HOME /root ENV OLDPWD / ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30250 30251 30252 30253 30254 30255 30256 30257 30258 30259 30260 30261 30262 30263 30264 30265 30266 30267 30268 30269 30270 30271 30272 30273 30274 30275 30276 30277 30278 30279 30280 30281 30282 30283 30284 30285 30286 30287 30288 30289 30290 30291 30292 30293 30294 30295 30296 30297 30298 30299 ENV WORKER_FREE_PORTS 30250 30251 30252 30253 30254 30255 30256 30257 30258 30259 30260 30261 30262 30263 30264 30265 30266 30267 30268 30269 30270 30271 30272 30273 30274 30275 30276 30277 30278 30279 30280 30281 30282 30283 30284 30285 30286 30287 30288 30289 30290 30291 30292 30293 30294 30295 30296 30297 30298 30299 ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_WORKER gw5 ENV PYTEST_XDIST_WORKER gw5 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference (setup) ENV PYTEST_CURRENT_TEST test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference (setup) CLUSTER INIT base_config_dir:/clickhouse-config CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Setup Keeper Setup Keeper Cluster name: project_name:roottests3cluster-gw5. Added instance name:s0_0_0 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env', '--project-name', 'roottests3cluster-gw5', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:roottests3cluster-gw5. Added instance name:s0_0_0 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env', '--project-name', 'roottests3cluster-gw5', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottests3cluster-gw5. Added instance name:s0_0_1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env', '--project-name', 'roottests3cluster-gw5', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:roottests3cluster-gw5. Added instance name:s0_0_1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env', '--project-name', 'roottests3cluster-gw5', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottests3cluster-gw5. Added instance name:s0_1_0 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env', '--project-name', 'roottests3cluster-gw5', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:roottests3cluster-gw5. Added instance name:s0_1_0 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env', '--project-name', 'roottests3cluster-gw5', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Starting cluster... Starting cluster... Running tests in /ClickHouse/tests/integration/test_s3_cluster/test.py Running tests in /ClickHouse/tests/integration/test_s3_cluster/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottestsshkeysauthentication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsshkeysauthentication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemstartstoplisten-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemstartstoplisten-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3zerocopyreplication-gw0 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3zerocopyreplication-gw0 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageredis-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageredis-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageazureblobstoragecluster-gw2 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageazureblobstoragecluster-gw2 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemqueries-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemqueries-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurl-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3cluster-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurl-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3cluster-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemmerges-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemmerges-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsshkeysauthentication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsshkeysauthentication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemstartstoplisten-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemstartstoplisten-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageredis-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageredis-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3zerocopyreplication-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3zerocopyreplication-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageazureblobstoragecluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageazureblobstoragecluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemqueries-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurl-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurl-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemqueries-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3cluster-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3cluster-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemmerges-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemmerges-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottests3zerocopyreplication-gw0 are DRIVER VOLUME NAME Docker volumes for project roottests3zerocopyreplication-gw0 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project roottestsshkeysauthentication-gw8 are DRIVER VOLUME NAME Docker volumes for project roottestsshkeysauthentication-gw8 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project roottestsystemstartstoplisten-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsystemstartstoplisten-gw9 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project rootteststorageazureblobstoragecluster-gw2 are DRIVER VOLUME NAME Docker volumes for project rootteststorageazureblobstoragecluster-gw2 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project rootteststorageredis-gw1 are DRIVER VOLUME NAME Docker volumes for project rootteststorageredis-gw1 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project rootteststorageurl-gw4 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurl-gw4 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project roottestsystemmerges-gw6 are DRIVER VOLUME NAME Docker volumes for project roottestsystemmerges-gw6 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project roottestsystemqueries-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemqueries-gw7 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project roottests3cluster-gw5 are DRIVER VOLUME NAME Docker volumes for project roottests3cluster-gw5 are DRIVER VOLUME NAME Docker networks for project roottests3zerocopyreplication-gw0 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3zerocopyreplication-gw0 are NETWORK ID NAME DRIVER SCOPE Cleanup called Cleanup called Docker networks for project rootteststorageurl-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurl-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsshkeysauthentication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsshkeysauthentication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemmerges-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemmerges-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemstartstoplisten-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemstartstoplisten-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemqueries-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemqueries-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageazureblobstoragecluster-gw2 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageazureblobstoragecluster-gw2 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageredis-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageredis-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3cluster-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3cluster-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemmerges-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemmerges-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurl-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurl-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemstartstoplisten-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemqueries-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemqueries-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemstartstoplisten-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3zerocopyreplication-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3zerocopyreplication-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageazureblobstoragecluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageazureblobstoragecluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsshkeysauthentication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsshkeysauthentication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3cluster-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3cluster-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageredis-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageredis-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemmerges-gw6 are DRIVER VOLUME NAME Docker volumes for project roottestsystemmerges-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemmerges-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemmerges-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project rootteststorageurl-gw4 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurl-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageurl-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageurl-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestsystemqueries-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemqueries-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemqueries-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemqueries-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestsshkeysauthentication-gw8 are DRIVER VOLUME NAME Docker volumes for project roottestsystemstartstoplisten-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsshkeysauthentication-gw8 are DRIVER VOLUME NAME Docker volumes for project roottestsystemstartstoplisten-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemstartstoplisten-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsshkeysauthentication-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsshkeysauthentication-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottests3zerocopyreplication-gw0 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemstartstoplisten-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottests3zerocopyreplication-gw0 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottests3zerocopyreplication-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottests3zerocopyreplication-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project rootteststorageredis-gw1 are DRIVER VOLUME NAME Docker volumes for project rootteststorageredis-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageredis-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageredis-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottests3cluster-gw5 are DRIVER VOLUME NAME Docker volumes for project roottests3cluster-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottests3cluster-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottests3cluster-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project rootteststorageazureblobstoragecluster-gw2 are DRIVER VOLUME NAME Docker volumes for project rootteststorageazureblobstoragecluster-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageazureblobstoragecluster-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageazureblobstoragecluster-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageurl-gw4 No running containers for project: rootteststorageurl-gw4 Trying to prune unused networks... Trying to prune unused networks... Unstopped containers: {} Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemqueries-gw7 No running containers for project: roottestsystemqueries-gw7 Unstopped containers: {} No running containers for project: roottestsystemmerges-gw6 Trying to prune unused networks... Trying to prune unused networks... No running containers for project: roottestsystemmerges-gw6 Trying to prune unused networks... Trying to prune unused networks... Unstopped containers: {} Unstopped containers: {} No running containers for project: roottests3zerocopyreplication-gw0 No running containers for project: roottests3zerocopyreplication-gw0 Trying to prune unused networks... Trying to prune unused networks... Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemstartstoplisten-gw9 No running containers for project: roottestsystemstartstoplisten-gw9 Trying to prune unused networks... Trying to prune unused networks... Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsshkeysauthentication-gw8 No running containers for project: roottestsshkeysauthentication-gw8 Trying to prune unused networks... Trying to prune unused networks... Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageredis-gw1 No running containers for project: rootteststorageredis-gw1 Trying to prune unused networks... Trying to prune unused networks... Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageazureblobstoragecluster-gw2 No running containers for project: rootteststorageazureblobstoragecluster-gw2 Trying to prune unused networks... Trying to prune unused networks... Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttablefunctionmongodblegacy-gw3 No running containers for project: roottesttablefunctionmongodblegacy-gw3 Trying to prune unused networks... Trying to prune unused networks... Unstopped containers: {} Unstopped containers: {} No running containers for project: roottests3cluster-gw5 No running containers for project: roottests3cluster-gw5 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Trying to prune unused images... Trying to prune unused images... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Command:[docker image prune -f] Command:[docker image prune -f] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Trying to prune unused images... Trying to prune unused images... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Command:[docker image prune -f] Command:[docker image prune -f] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Exitcode:1 Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Exitcode:1 Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Exitcode:1 Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Stderr:Error response from daemon: a prune operation is already running Stderr:Error response from daemon: a prune operation is already running Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Exitcode:1 Exitcode:1 Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Exitcode:1 Trying to prune unused volumes... Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Exitcode:1 Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:1 Stdout:1 Volumes pruned: 1 Volumes pruned: 1 Setup directory for instance: node1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_merges/configs/logs_config.xml'] to /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_merges/configs/logs_config.xml'] to /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/database Setup database dir /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/database Setup logs dir /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/logs Setup logs dir /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Setup directory for instance: node2 Create directory for configuration generated in this helper Stdout:1 Stdout:1 Create directory for configuration generated in this helper Create directory for common tests configuration Volumes pruned: 1 Volumes pruned: 1 Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Setup directory for instance: node1 Setup directory for instance: node1 Generate and write macros file Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Copy custom test config files ['/ClickHouse/tests/integration/test_system_merges/configs/logs_config.xml'] to /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/configs/config.d Images pruned Images pruned Copy custom test config files ['/ClickHouse/tests/integration/test_system_merges/configs/logs_config.xml'] to /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/configs/config.d Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Create directory for configuration generated in this helper Create directory for configuration generated in this helper Stdout:1 Stdout:1 Setup database dir /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/database Setup database dir /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/database Setup logs dir /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/logs Stdout:1 Setup logs dir /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/logs Stdout:1 Volumes pruned: 1 Volumes pruned: 1 Volumes pruned: 1 Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Volumes pruned: 1 Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: ch1 Setup directory for instance: ch1 Setup directory for instance: node1 Setup directory for instance: node1 Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Create directory for common tests configuration Create directory for common tests configuration Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Copy common configuration from helpers Copy common configuration from helpers No config file found No config file found Copy common configuration from helpers Copy common configuration from helpers Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_s3_zero_copy_replication/configs/config.d/s3.xml'] to /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_s3_zero_copy_replication/configs/config.d/s3.xml'] to /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/configs/config.d Generate and write macros file Generate and write macros file Setup database dir /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/database Setup database dir /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/database Setup logs dir /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/logs Copy custom test config files ['/ClickHouse/tests/integration/test_system_queries/configs/config.d/clusters_config.xml', '/ClickHouse/tests/integration/test_system_queries/configs/config.d/query_log.xml'] to /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/configs/config.d Setup logs dir /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/logs Copy custom test config files ['/ClickHouse/tests/integration/test_system_queries/configs/config.d/clusters_config.xml', '/ClickHouse/tests/integration/test_system_queries/configs/config.d/query_log.xml'] to /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/configs/config.d Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_s3_zero_copy_replication/configs/config.d/s3.xml'] to /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/database Setup database dir /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/database Copy custom test config files ['/ClickHouse/tests/integration/test_s3_zero_copy_replication/configs/config.d/s3.xml'] to /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/configs/config.d Setup logs dir /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/logs Setup logs dir /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup database dir /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/database Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env Setup database dir /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/database Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env Setup logs dir /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/logs Setup logs dir /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/coordination', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/coordination', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Stdout:1 Stdout:1 Volumes pruned: 1 Stdout:1 Stdout:1 Stdout:1 Stdout:1 Volumes pruned: 1 Volumes pruned: 1 Volumes pruned: 1 Volumes pruned: 1 Setup directory for instance: node Setup directory for instance: node Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Generate and write macros file Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/configs/config.d Copy custom test config files [] to /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/database Setup database dir /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/database Copy custom test config files [] to /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/configs/config.d Copy custom test config files [] to /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/configs/config.d Setup logs dir /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/logs Setup logs dir /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/logs http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Create directory for common tests configuration Create directory for common tests configuration Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'REDIS_HOST': 'redis1', 'REDIS_EXTERNAL_PORT': '30050', 'REDIS_INTERNAL_PORT': '6379'} stored in /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'REDIS_HOST': 'redis1', 'REDIS_EXTERNAL_PORT': '30050', 'REDIS_INTERNAL_PORT': '6379'} stored in /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env Setup database dir /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/database Copy common configuration from helpers Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/database Setup logs dir /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/logs Setup logs dir /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/logs Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml pull] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env Generate and write macros file Generate and write macros file Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env Stdout:1 Stdout:1 Volumes pruned: 1 Volumes pruned: 1 Copy custom test config files ['/ClickHouse/tests/integration/test_storage_url/configs/conf.xml', '/ClickHouse/tests/integration/test_storage_url/configs/named_collections.xml'] to /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_storage_url/configs/conf.xml', '/ClickHouse/tests/integration/test_storage_url/configs/named_collections.xml'] to /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/configs/config.d Setup directory for instance: node_0 Setup directory for instance: node_0 Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Create directory for configuration generated in this helper Create directory for configuration generated in this helper No config file found No config file found Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Volumes pruned: 1 Setup directory for instance: s0_0_0 Setup directory for instance: s0_0_0 Copy custom test config files ['/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/named_collections.xml', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/cluster.xml'] to /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/named_collections.xml', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/cluster.xml'] to /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/configs/config.d Setup database dir /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/database Setup database dir /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/database Setup logs dir /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/logs Setup logs dir /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/logs Stdout:1 Stdout:1 Create directory for configuration generated in this helper Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Create directory for configuration generated in this helper Volumes pruned: 1 Volumes pruned: 1 Create directory for common tests configuration Create directory for common tests configuration Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'NGINX_EXTERNAL_PORT': '30201'} stored in /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'NGINX_EXTERNAL_PORT': '30201'} stored in /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env Setup directory for instance: main_node Setup directory for instance: main_node Copy common configuration from helpers Copy common configuration from helpers Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Generate and write macros file Generate and write macros file Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Create directory for configuration generated in this helper No config file found Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Copy custom test config files ['/ClickHouse/tests/integration/test_s3_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_s3_cluster/configs/named_collections.xml'] to /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_s3_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_s3_cluster/configs/named_collections.xml'] to /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/configs/config.d Setup database dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/database Setup database dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/database Setup logs dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/logs Setup logs dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/logs Setup database dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/database Setup database dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/database Setup logs dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/logs Setup logs dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Generate and write macros file Generate and write macros file Setup directory for instance: s0_0_1 Setup directory for instance: s0_0_1 Setup directory for instance: node_1 Setup directory for instance: node_1 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Copy custom test config files ['/ClickHouse/tests/integration/test_system_start_stop_listen/configs/cluster.xml', '/ClickHouse/tests/integration/test_system_start_stop_listen/configs/protocols.xml', '/ClickHouse/tests/integration/test_system_start_stop_listen/configs/server.crt', '/ClickHouse/tests/integration/test_system_start_stop_listen/configs/server.key'] to /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_start_stop_listen/configs/cluster.xml', '/ClickHouse/tests/integration/test_system_start_stop_listen/configs/protocols.xml', '/ClickHouse/tests/integration/test_system_start_stop_listen/configs/server.crt', '/ClickHouse/tests/integration/test_system_start_stop_listen/configs/server.key'] to /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/configs/config.d Create directory for common tests configuration Create directory for common tests configuration Create directory for configuration generated in this helper Copy common configuration from helpers Copy common configuration from helpers Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Generate and write macros file Generate and write macros file Copy common configuration from helpers Copy common configuration from helpers Copy custom test config files ['/ClickHouse/tests/integration/test_s3_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_s3_cluster/configs/named_collections.xml'] to /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_s3_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_s3_cluster/configs/named_collections.xml'] to /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/configs/config.d Stdout:1 Stdout:1 Setup database dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/database Setup database dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/database Setup database dir /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/database Setup database dir /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/database Generate and write macros file Generate and write macros file Volumes pruned: 1 Volumes pruned: 1 Setup logs dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/logs Setup logs dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/logs Setup directory for instance: node Setup logs dir /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/logs Setup directory for instance: node Setup logs dir /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: s0_1_0 Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Copy custom test config files ['/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/named_collections.xml', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/cluster.xml'] to /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/configs/config.d Setup directory for instance: s0_1_0 Copy custom test config files ['/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/named_collections.xml', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/cluster.xml'] to /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/configs/config.d Setup directory for instance: backup_node Setup directory for instance: backup_node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Create directory for common tests configuration Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/database Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/database Copy common configuration from helpers Create directory for configuration generated in this helper Create directory for configuration generated in this helper http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Setup logs dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/logs Setup logs dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/logs Create directory for common tests configuration Create directory for common tests configuration Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Generate and write macros file Generate and write macros file Generate and write macros file Generate and write macros file Copy common configuration from helpers Copy common configuration from helpers Setup directory for instance: node_2 Setup directory for instance: node_2 Copy custom test config files ['/ClickHouse/tests/integration/test_s3_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_s3_cluster/configs/named_collections.xml'] to /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_s3_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_s3_cluster/configs/named_collections.xml'] to /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/configs/config.d Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml pull] Generate and write macros file Generate and write macros file Create directory for configuration generated in this helper Create directory for configuration generated in this helper Setup database dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/database Setup database dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/database Create directory for common tests configuration Create directory for common tests configuration Setup logs dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/logs Setup logs dir /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/logs Copy custom test config files ['/ClickHouse/tests/integration/test_system_start_stop_listen/configs/cluster.xml'] to /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_start_stop_listen/configs/cluster.xml'] to /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/configs/config.d Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Copy common configuration from helpers Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/database Setup database dir /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/database Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/coordination', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/coordination', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env Setup logs dir /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/logs Setup logs dir /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env Generate and write macros file Generate and write macros file Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Copy custom test config files ['/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/named_collections.xml', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/cluster.xml'] to /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/named_collections.xml', '/ClickHouse/tests/integration/test_storage_azure_blob_storage/configs/cluster.xml'] to /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_table_function_mongodb_legacy/mongo_secure_config/config.d/ssl_conf.xml', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/configs/feature_flag.xml'] to /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_table_function_mongodb_legacy/mongo_secure_config/config.d/ssl_conf.xml', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/configs/feature_flag.xml'] to /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/configs/config.d No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Command:[docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml pull] Setup database dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/database Setup database dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/database Setup logs dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/logs Setup logs dir /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/logs http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Setup database dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/database Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup database dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/database Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup logs dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs Setup logs dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'AZURITE_PORT': '30102', 'AZURITE_STORAGE_ACCOUNT_URL': 'http://azurite1:30102/devstoreaccount1', 'AZURITE_CONNECTION_STRING': 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30102/devstoreaccount1;'} stored in /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'AZURITE_PORT': '30102', 'AZURITE_STORAGE_ACCOUNT_URL': 'http://azurite1:30102/devstoreaccount1', 'AZURITE_CONNECTION_STRING': 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30102/devstoreaccount1;'} stored in /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml pull] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MONGO_HOST': 'mongo1', 'MONGO_EXTERNAL_PORT': '30152', 'MONGO_INTERNAL_PORT': '27017', 'MONGO_NO_CRED_EXTERNAL_PORT': '30153', 'MONGO_NO_CRED_INTERNAL_PORT': '27017', 'MONGO_SECURE_EXTERNAL_PORT': '30154', 'MONGO_SECURE_INTERNAL_PORT': '27017', 'MONGO_SECURE_CONFIG_DIR': '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/mongo_secure_config'} stored in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MONGO_HOST': 'mongo1', 'MONGO_EXTERNAL_PORT': '30152', 'MONGO_INTERNAL_PORT': '27017', 'MONGO_NO_CRED_EXTERNAL_PORT': '30153', 'MONGO_NO_CRED_INTERNAL_PORT': '27017', 'MONGO_SECURE_EXTERNAL_PORT': '30154', 'MONGO_SECURE_INTERNAL_PORT': '27017', 'MONGO_SECURE_CONFIG_DIR': '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/mongo_secure_config'} stored in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found No config file found No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml pull] http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml pull] http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml pull] http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml pull] http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml pull] Stderr: node Pulling Stderr: node Pulling Stderr: node Pulled Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate] Stderr: node1 Skipped - Image is already being pulled by zoo2 Stderr: node1 Skipped - Image is already being pulled by zoo2 Stderr: zoo3 Skipped - Image is already being pulled by zoo2 Stderr: zoo3 Skipped - Image is already being pulled by zoo2 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: node2 Skipped - Image is already being pulled by zoo2 Stderr: node2 Skipped - Image is already being pulled by zoo2 Stderr: zoo1 Skipped - Image is already being pulled by zoo2 Stderr: zoo1 Skipped - Image is already being pulled by zoo2 Stderr: zoo2 Pulling Stderr: zoo2 Pulling Stderr: resolver Pulling Stderr: resolver Pulling Stderr: proxy1 Pulling Stderr: proxy1 Pulling Stderr: minio1 Pulling Stderr: minio1 Pulling Stderr: proxy1 Pulled Stderr: proxy1 Pulled Stderr: zoo2 Pulled Stderr: zoo2 Pulled Stderr: resolver Pulled Stderr: resolver Pulled Stderr: minio1 Pulled Stderr: minio1 Pulled Setup ZooKeeper Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/log', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/config', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/coordination', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/log', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/config', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/coordination', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/log', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/config', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/coordination'] Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/log', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/config', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper1/coordination', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/log', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/config', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper2/coordination', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/log', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/config', '/ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/keeper3/coordination'] Command:[docker compose --project-name roottests3zerocopyreplication-gw0 --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Command:[docker compose --project-name roottests3zerocopyreplication-gw0 --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: node2 Skipped - Image is already being pulled by zoo2 Stderr: node2 Skipped - Image is already being pulled by zoo2 Stderr: node1 Skipped - Image is already being pulled by zoo2 Stderr: node1 Skipped - Image is already being pulled by zoo2 Stderr: zoo3 Skipped - Image is already being pulled by zoo2 Stderr: zoo3 Skipped - Image is already being pulled by zoo2 Stderr: zoo1 Skipped - Image is already being pulled by zoo2 Stderr: zoo1 Skipped - Image is already being pulled by zoo2 Stderr: zoo2 Pulling Stderr: zoo2 Pulling Stderr: zoo2 Pulled Stderr: zoo2 Pulled Setup ZooKeeper Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/coordination'] Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/keeper3/coordination'] Command:[docker compose --project-name roottestsystemmerges-gw6 --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Command:[docker compose --project-name roottestsystemmerges-gw6 --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: ch1 Pulling Stderr: ch1 Pulling Stderr: ch1 Pulled Stderr: ch1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml up -d --no-recreate] Stderr: s0_1_0 Skipped - Image is already being pulled by s0_0_0 Stderr: s0_1_0 Skipped - Image is already being pulled by s0_0_0 Stderr: zoo1 Skipped - Image is already being pulled by s0_0_0 Stderr: zoo1 Skipped - Image is already being pulled by s0_0_0 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: s0_0_1 Skipped - Image is already being pulled by s0_0_0 Stderr: s0_0_1 Skipped - Image is already being pulled by s0_0_0 Stderr: proxy1 Pulling Stderr: proxy1 Pulling Stderr: zoo3 Skipped - Image is already being pulled by s0_0_0 Stderr: zoo3 Skipped - Image is already being pulled by s0_0_0 Stderr: minio1 Pulling Stderr: minio1 Pulling Stderr: zoo2 Skipped - Image is already being pulled by s0_0_0 Stderr: zoo2 Skipped - Image is already being pulled by s0_0_0 Stderr: resolver Pulling Stderr: resolver Pulling Stderr: s0_0_0 Pulling Stderr: s0_0_0 Pulling Stderr: minio1 Pulled Stderr: minio1 Pulled Stderr: proxy1 Pulled Stderr: proxy1 Pulled Stderr: resolver Pulled Stderr: resolver Pulled Stderr: s0_0_0 Pulled Stderr: s0_0_0 Pulled Setup ZooKeeper Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/log', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/config', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/coordination', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/log', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/config', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/coordination', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/log', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/config', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/coordination'] Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/log', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/config', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper1/coordination', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/log', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/config', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper2/coordination', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/log', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/config', '/ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/keeper3/coordination'] Command:[docker compose --project-name roottests3cluster-gw5 --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Command:[docker compose --project-name roottests3cluster-gw5 --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: main_node Skipped - Image is already being pulled by backup_node Stderr: main_node Skipped - Image is already being pulled by backup_node Stderr: zoo3 Skipped - Image is already being pulled by backup_node Stderr: zoo3 Skipped - Image is already being pulled by backup_node Stderr: zoo1 Skipped - Image is already being pulled by backup_node Stderr: zoo1 Skipped - Image is already being pulled by backup_node Stderr: zoo2 Skipped - Image is already being pulled by backup_node Stderr: zoo2 Skipped - Image is already being pulled by backup_node Stderr: backup_node Pulling Stderr: backup_node Pulling Stderr: backup_node Pulled Stderr: backup_node Pulled Setup ZooKeeper Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/log', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/config', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/coordination', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/log', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/config', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/coordination', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/log', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/config', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/coordination'] Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/log', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/config', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper1/coordination', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/log', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/config', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper2/coordination', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/log', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/config', '/ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/keeper3/coordination'] Command:[docker compose --project-name roottestsystemstartstoplisten-gw9 --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Command:[docker compose --project-name roottestsystemstartstoplisten-gw9 --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: Network roottestsshkeysauthentication-gw8_default Creating Stderr: Network roottestsshkeysauthentication-gw8_default Creating Stderr: Network roottestsshkeysauthentication-gw8_default Created Stderr: Network roottestsshkeysauthentication-gw8_default Created Stderr: Container roottestsshkeysauthentication-gw8-node-1 Creating Stderr: Container roottestsshkeysauthentication-gw8-node-1 Creating Stderr: Container roottestsshkeysauthentication-gw8-node-1 Created Stderr: Container roottestsshkeysauthentication-gw8-node-1 Created Stderr: Container roottestsshkeysauthentication-gw8-node-1 Starting Stderr: Container roottestsshkeysauthentication-gw8-node-1 Starting Stderr: Container roottestsshkeysauthentication-gw8-node-1 Started Stderr: Container roottestsshkeysauthentication-gw8-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestsshkeysauthentication-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsshkeysauthentication-gw8-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestsshkeysauthentication-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsshkeysauthentication-gw8-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.2... Waiting for ClickHouse start in node, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestsshkeysauthentication-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsshkeysauthentication-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None Stderr: nginx Pulling Stderr: nginx Pulling Stderr: node1 Pulling Stderr: node1 Pulling Stderr: node1 Pulled Stderr: node1 Pulled Stderr: 4abcf2066143 Already exists Stderr: 4abcf2066143 Already exists Stderr: fc21a1d387f5 Pulling fs layer Stderr: fc21a1d387f5 Pulling fs layer Stderr: e6ef242c1570 Pulling fs layer Stderr: e6ef242c1570 Pulling fs layer Stderr: 13fcfbc94648 Pulling fs layer Stderr: 13fcfbc94648 Pulling fs layer Stderr: d4bca490e609 Pulling fs layer Stderr: d4bca490e609 Pulling fs layer Stderr: 5406ed7b06d9 Pulling fs layer Stderr: 5406ed7b06d9 Pulling fs layer Stderr: 8a3742a9529d Pulling fs layer Stderr: 8a3742a9529d Pulling fs layer Stderr: 1c5d1d091cc8 Pulling fs layer Stderr: 1c5d1d091cc8 Pulling fs layer Stderr: 82fc7741492f Pulling fs layer Stderr: 82fc7741492f Pulling fs layer Stderr: 5406ed7b06d9 Waiting Stderr: 5406ed7b06d9 Waiting Stderr: 8a3742a9529d Waiting Stderr: 8a3742a9529d Waiting Stderr: 82fc7741492f Waiting Stderr: 82fc7741492f Waiting Stderr: 1c5d1d091cc8 Waiting Stderr: 1c5d1d091cc8 Waiting Stderr: d4bca490e609 Waiting Stderr: d4bca490e609 Waiting Stderr: 13fcfbc94648 Downloading [==================================================>] 954B/954B Stderr: 13fcfbc94648 Downloading [==================================================>] 954B/954B Stderr: 13fcfbc94648 Verifying Checksum Stderr: 13fcfbc94648 Verifying Checksum Stderr: 13fcfbc94648 Download complete Stderr: 13fcfbc94648 Download complete Stderr: e6ef242c1570 Downloading [==================================================>] 629B/629B Stderr: e6ef242c1570 Downloading [==================================================>] 629B/629B Stderr: e6ef242c1570 Verifying Checksum Stderr: e6ef242c1570 Verifying Checksum Stderr: e6ef242c1570 Download complete Stderr: e6ef242c1570 Download complete Stderr: 5406ed7b06d9 Downloading [==================================================>] 1.209kB/1.209kB Stderr: 5406ed7b06d9 Downloading [==================================================>] 1.209kB/1.209kB Stderr: d4bca490e609 Downloading [==================================================>] 393B/393B Stderr: d4bca490e609 Downloading [==================================================>] 393B/393B Stderr: d4bca490e609 Verifying Checksum Stderr: d4bca490e609 Verifying Checksum Stderr: d4bca490e609 Download complete Stderr: d4bca490e609 Download complete Stderr: 8a3742a9529d Download complete Stderr: 8a3742a9529d Download complete Stderr: 1c5d1d091cc8 Downloading [==================================================>] 505B/505B Stderr: 1c5d1d091cc8 Downloading [==================================================>] 505B/505B Stderr: 1c5d1d091cc8 Verifying Checksum Stderr: 1c5d1d091cc8 Verifying Checksum Stderr: 1c5d1d091cc8 Download complete Stderr: 1c5d1d091cc8 Download complete Stderr: fc21a1d387f5 Downloading [> ] 42.74kB/3.985MB Stderr: fc21a1d387f5 Downloading [> ] 42.74kB/3.985MB Stderr: 82fc7741492f Downloading [==================================================>] 181B/181B Stderr: 82fc7741492f Downloading [==================================================>] 181B/181B Stderr: 82fc7741492f Download complete Stderr: 82fc7741492f Download complete Stderr: fc21a1d387f5 Verifying Checksum Stderr: fc21a1d387f5 Verifying Checksum Stderr: fc21a1d387f5 Download complete Stderr: fc21a1d387f5 Download complete Stderr: fc21a1d387f5 Extracting [> ] 65.54kB/3.985MB Stderr: fc21a1d387f5 Extracting [> ] 65.54kB/3.985MB Stderr: fc21a1d387f5 Extracting [=================================================> ] 3.932MB/3.985MB Stderr: fc21a1d387f5 Extracting [=================================================> ] 3.932MB/3.985MB Stderr: fc21a1d387f5 Extracting [==================================================>] 3.985MB/3.985MB Stderr: fc21a1d387f5 Extracting [==================================================>] 3.985MB/3.985MB Stderr: fc21a1d387f5 Extracting [==================================================>] 3.985MB/3.985MB Stderr: fc21a1d387f5 Extracting [==================================================>] 3.985MB/3.985MB Stderr: fc21a1d387f5 Pull complete Stderr: fc21a1d387f5 Pull complete Stderr: e6ef242c1570 Extracting [==================================================>] 629B/629B Stderr: e6ef242c1570 Extracting [==================================================>] 629B/629B Stderr: e6ef242c1570 Extracting [==================================================>] 629B/629B Stderr: e6ef242c1570 Extracting [==================================================>] 629B/629B Stderr: e6ef242c1570 Pull complete Stderr: e6ef242c1570 Pull complete Stderr: 13fcfbc94648 Extracting [==================================================>] 954B/954B Stderr: 13fcfbc94648 Extracting [==================================================>] 954B/954B Stderr: 13fcfbc94648 Extracting [==================================================>] 954B/954B Stderr: 13fcfbc94648 Extracting [==================================================>] 954B/954B Stderr: 13fcfbc94648 Pull complete Stderr: 13fcfbc94648 Pull complete Stderr: d4bca490e609 Extracting [==================================================>] 393B/393B Stderr: d4bca490e609 Extracting [==================================================>] 393B/393B Stderr: d4bca490e609 Extracting [==================================================>] 393B/393B Stderr: d4bca490e609 Extracting [==================================================>] 393B/393B Stderr: d4bca490e609 Pull complete Stderr: d4bca490e609 Pull complete Stderr: 5406ed7b06d9 Extracting [==================================================>] 1.209kB/1.209kB Stderr: 5406ed7b06d9 Extracting [==================================================>] 1.209kB/1.209kB Stderr: 5406ed7b06d9 Extracting [==================================================>] 1.209kB/1.209kB Stderr: 5406ed7b06d9 Extracting [==================================================>] 1.209kB/1.209kB Stderr: 5406ed7b06d9 Pull complete Stderr: 5406ed7b06d9 Pull complete Stderr: 8a3742a9529d Extracting [==================================================>] 1.398kB/1.398kB Stderr: 8a3742a9529d Extracting [==================================================>] 1.398kB/1.398kB Stderr: 8a3742a9529d Extracting [==================================================>] 1.398kB/1.398kB Stderr: 8a3742a9529d Extracting [==================================================>] 1.398kB/1.398kB Stderr: 8a3742a9529d Pull complete Stderr: 8a3742a9529d Pull complete Stderr: 1c5d1d091cc8 Extracting [==================================================>] 505B/505B Stderr: 1c5d1d091cc8 Extracting [==================================================>] 505B/505B Stderr: 1c5d1d091cc8 Extracting [==================================================>] 505B/505B Stderr: 1c5d1d091cc8 Extracting [==================================================>] 505B/505B Stderr: 1c5d1d091cc8 Pull complete Stderr: 1c5d1d091cc8 Pull complete Stderr: 82fc7741492f Extracting [==================================================>] 181B/181B Stderr: 82fc7741492f Extracting [==================================================>] 181B/181B Stderr: 82fc7741492f Extracting [==================================================>] 181B/181B Stderr: 82fc7741492f Extracting [==================================================>] 181B/181B Stderr: 82fc7741492f Pull complete Stderr: 82fc7741492f Pull complete Stderr: nginx Pulled Stderr: nginx Pulled Setup nginx Setup nginx Command:[docker compose --project-name rootteststorageurl-gw4 --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml --verbose up -d --renew-anon-volumes] Command:[docker compose --project-name rootteststorageurl-gw4 --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml --verbose up -d --renew-anon-volumes] Stderr: node Pulling Stderr: node Pulling Stderr: redis1 Pulling Stderr: redis1 Pulling Stderr: 6e909acdb790 Already exists Stderr: 6e909acdb790 Already exists Stderr: d9b98f603d6b Pulling fs layer Stderr: d9b98f603d6b Pulling fs layer Stderr: 513666b05c81 Pulling fs layer Stderr: 513666b05c81 Pulling fs layer Stderr: 764e7fbfdb05 Pulling fs layer Stderr: 764e7fbfdb05 Pulling fs layer Stderr: 2884ccae34c1 Pulling fs layer Stderr: 2884ccae34c1 Pulling fs layer Stderr: df4660a56a60 Pulling fs layer Stderr: df4660a56a60 Pulling fs layer Stderr: 4f4fb700ef54 Pulling fs layer Stderr: 4f4fb700ef54 Pulling fs layer Stderr: b138a3179b82 Pulling fs layer Stderr: b138a3179b82 Pulling fs layer Stderr: 2884ccae34c1 Waiting Stderr: 2884ccae34c1 Waiting Stderr: df4660a56a60 Waiting Stderr: df4660a56a60 Waiting Stderr: b138a3179b82 Waiting Stderr: b138a3179b82 Waiting Stderr: 4f4fb700ef54 Waiting Stderr: 4f4fb700ef54 Waiting Stderr: 513666b05c81 Downloading [==================================================>] 873B/873B Stderr: 513666b05c81 Downloading [==================================================>] 873B/873B Stderr: 513666b05c81 Verifying Checksum Stderr: 513666b05c81 Verifying Checksum Stderr: 513666b05c81 Download complete Stderr: 513666b05c81 Download complete Stderr: d9b98f603d6b Downloading [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Downloading [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Download complete Stderr: d9b98f603d6b Download complete Stderr: d9b98f603d6b Extracting [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Extracting [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Extracting [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Extracting [==================================================>] 1.102kB/1.102kB Stderr: 764e7fbfdb05 Downloading [> ] 22.47kB/1.438MB Stderr: 764e7fbfdb05 Downloading [> ] 22.47kB/1.438MB Stderr: 2884ccae34c1 Downloading [> ] 169.6kB/15.33MB Stderr: 2884ccae34c1 Downloading [> ] 169.6kB/15.33MB Stderr: 764e7fbfdb05 Verifying Checksum Stderr: 764e7fbfdb05 Verifying Checksum Stderr: 764e7fbfdb05 Download complete Stderr: 764e7fbfdb05 Download complete Stderr: d9b98f603d6b Pull complete Stderr: d9b98f603d6b Pull complete Stderr: df4660a56a60 Downloading [==================================================>] 97B/97B Stderr: df4660a56a60 Downloading [==================================================>] 97B/97B Stderr: df4660a56a60 Verifying Checksum Stderr: df4660a56a60 Verifying Checksum Stderr: df4660a56a60 Download complete Stderr: df4660a56a60 Download complete Stderr: 513666b05c81 Extracting [==================================================>] 873B/873B Stderr: 513666b05c81 Extracting [==================================================>] 873B/873B Stderr: 513666b05c81 Extracting [==================================================>] 873B/873B Stderr: 513666b05c81 Extracting [==================================================>] 873B/873B Stderr: 4f4fb700ef54 Downloading [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Downloading [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Download complete Stderr: 4f4fb700ef54 Download complete Stderr: node Pulled Stderr: node Pulled Stderr: b138a3179b82 Downloading [==================================================>] 572B/572B Stderr: b138a3179b82 Downloading [==================================================>] 572B/572B Stderr: b138a3179b82 Verifying Checksum Stderr: b138a3179b82 Verifying Checksum Stderr: b138a3179b82 Download complete Stderr: b138a3179b82 Download complete Stderr: 513666b05c81 Pull complete Stderr: 513666b05c81 Pull complete Stderr: 764e7fbfdb05 Extracting [=> ] 32.77kB/1.438MB Stderr: 764e7fbfdb05 Extracting [=> ] 32.77kB/1.438MB Stderr: 2884ccae34c1 Downloading [==================================================>] 15.33MB/15.33MB Stderr: 2884ccae34c1 Downloading [==================================================>] 15.33MB/15.33MB Stderr: 2884ccae34c1 Verifying Checksum Stderr: 2884ccae34c1 Verifying Checksum Stderr: 2884ccae34c1 Download complete Stderr: 2884ccae34c1 Download complete Stderr: 764e7fbfdb05 Extracting [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Extracting [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Extracting [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Extracting [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Pull complete Stderr: 764e7fbfdb05 Pull complete Stderr: 2884ccae34c1 Extracting [> ] 163.8kB/15.33MB Stderr: 2884ccae34c1 Extracting [> ] 163.8kB/15.33MB Stderr: 2884ccae34c1 Extracting [========================> ] 7.373MB/15.33MB Stderr: 2884ccae34c1 Extracting [========================> ] 7.373MB/15.33MB Stderr: 2884ccae34c1 Extracting [==================================================>] 15.33MB/15.33MB Stderr: 2884ccae34c1 Extracting [==================================================>] 15.33MB/15.33MB Stderr: 2884ccae34c1 Pull complete Stderr: 2884ccae34c1 Pull complete Stderr: df4660a56a60 Extracting [==================================================>] 97B/97B Stderr: df4660a56a60 Extracting [==================================================>] 97B/97B Stderr: df4660a56a60 Extracting [==================================================>] 97B/97B Stderr: df4660a56a60 Extracting [==================================================>] 97B/97B Stderr: df4660a56a60 Pull complete Stderr: df4660a56a60 Pull complete Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Pull complete Stderr: 4f4fb700ef54 Pull complete Stderr: b138a3179b82 Extracting [==================================================>] 572B/572B Stderr: b138a3179b82 Extracting [==================================================>] 572B/572B Stderr: b138a3179b82 Extracting [==================================================>] 572B/572B Stderr: b138a3179b82 Extracting [==================================================>] 572B/572B Stderr: b138a3179b82 Pull complete Stderr: b138a3179b82 Pull complete Stderr: redis1 Pulled Stderr: redis1 Pulled Setup Redis Setup Redis Command:[docker compose --project-name rootteststorageredis-gw1 --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml --verbose up -d] Command:[docker compose --project-name rootteststorageredis-gw1 --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml --verbose up -d] http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None Stderr: Network roottestsystemqueries-gw7_default Creating Stderr: Network roottestsystemqueries-gw7_default Creating Stderr: Network roottestsystemqueries-gw7_default Created Stderr: Network roottestsystemqueries-gw7_default Created Stderr: Container roottestsystemqueries-gw7-ch1-1 Creating Stderr: Container roottestsystemqueries-gw7-ch1-1 Creating Stderr: Container roottestsystemqueries-gw7-ch1-1 Created Stderr: Container roottestsystemqueries-gw7-ch1-1 Created Stderr: Container roottestsystemqueries-gw7-ch1-1 Starting Stderr: Container roottestsystemqueries-gw7-ch1-1 Starting Stderr: Container roottestsystemqueries-gw7-ch1-1 Started Stderr: Container roottestsystemqueries-gw7-ch1-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=ch1 get_instance_ip instance_name=ch1 http://localhost:None "GET /v1.46/containers/roottestsystemqueries-gw7-ch1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemqueries-gw7-ch1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=ch1 get_instance_ip instance_name=ch1 http://localhost:None "GET /v1.46/containers/roottestsystemqueries-gw7-ch1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemqueries-gw7-ch1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in ch1, ip: 172.16.2.2... Waiting for ClickHouse start in ch1, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottestsystemqueries-gw7-ch1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemqueries-gw7-ch1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4dc9347011a9a864285de69d110f79e98b707761cc7ca34c80a8ef11478df467/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Executing query SELECT currentUser() on node Executing query SELECT currentUser() on node http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottests3cluster-gw5_default Creating Stderr: Network roottests3cluster-gw5_default Creating Stderr: Network roottests3cluster-gw5_default Created Stderr: Network roottests3cluster-gw5_default Created Stderr: Container roottests3cluster-gw5-zoo3-1 Creating Stderr: Container roottests3cluster-gw5-zoo3-1 Creating Stderr: Container roottests3cluster-gw5-zoo1-1 Creating Stderr: Container roottests3cluster-gw5-zoo1-1 Creating Stderr: Container roottests3cluster-gw5-zoo2-1 Creating Stderr: Container roottests3cluster-gw5-zoo2-1 Creating Stderr: Container roottests3cluster-gw5-zoo2-1 Created Stderr: Container roottests3cluster-gw5-zoo2-1 Created Stderr: Container roottests3cluster-gw5-zoo3-1 Created Stderr: Container roottests3cluster-gw5-zoo3-1 Created Stderr: Container roottests3cluster-gw5-zoo1-1 Created Stderr: Container roottests3cluster-gw5-zoo1-1 Created Stderr: Container roottests3cluster-gw5-zoo1-1 Starting Stderr: Container roottests3cluster-gw5-zoo1-1 Starting Stderr: Container roottests3cluster-gw5-zoo2-1 Starting Stderr: Container roottests3cluster-gw5-zoo2-1 Starting Stderr: Container roottests3cluster-gw5-zoo3-1 Starting Stderr: Container roottests3cluster-gw5-zoo3-1 Starting Stderr: Container roottests3cluster-gw5-zoo1-1 Started Stderr: Container roottests3cluster-gw5-zoo1-1 Started Stderr: Container roottests3cluster-gw5-zoo3-1 Started Stderr: Container roottests3cluster-gw5-zoo3-1 Started Stderr: Container roottests3cluster-gw5-zoo2-1 Started Stderr: Container roottests3cluster-gw5-zoo2-1 Started Stderr:time="2025-04-02T04:14:21Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:21Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:21Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:21Z" level=debug msg="otel error" error="" Wait ZooKeeper to start Wait ZooKeeper to start get_instance_ip instance_name=zoo1 get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-zoo1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.3.4, port:2181, use_ssl:False get_kazoo_client: zoo1, ip:172.16.3.4, port:2181, use_ssl:False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused [gw8] PASSED test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 Executing query SELECT currentUser() on node Executing query SELECT currentUser() on node http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Stderr: Network roottestsystemmerges-gw6_default Creating Stderr: Network roottestsystemmerges-gw6_default Creating Stderr: Network roottestsystemmerges-gw6_default Created Stderr: Network roottestsystemmerges-gw6_default Created Stderr: Container roottestsystemmerges-gw6-zoo1-1 Creating Stderr: Container roottestsystemmerges-gw6-zoo1-1 Creating Stderr: Container roottestsystemmerges-gw6-zoo2-1 Creating Stderr: Container roottestsystemmerges-gw6-zoo2-1 Creating Stderr: Container roottestsystemmerges-gw6-zoo3-1 Creating Stderr: Container roottestsystemmerges-gw6-zoo3-1 Creating Stderr: Container roottestsystemmerges-gw6-zoo1-1 Created Stderr: Container roottestsystemmerges-gw6-zoo1-1 Created Stderr: Container roottestsystemmerges-gw6-zoo3-1 Created Stderr: Container roottestsystemmerges-gw6-zoo3-1 Created Stderr: Container roottestsystemmerges-gw6-zoo2-1 Created Stderr: Container roottestsystemmerges-gw6-zoo2-1 Created Stderr: Container roottestsystemmerges-gw6-zoo1-1 Starting Stderr: Container roottestsystemmerges-gw6-zoo1-1 Starting Stderr: Container roottestsystemmerges-gw6-zoo2-1 Starting Stderr: Container roottestsystemmerges-gw6-zoo2-1 Starting Stderr: Container roottestsystemmerges-gw6-zoo3-1 Starting Stderr: Container roottestsystemmerges-gw6-zoo3-1 Starting Stderr: Container roottestsystemmerges-gw6-zoo3-1 Started Stderr: Container roottestsystemmerges-gw6-zoo3-1 Started Stderr: Container roottestsystemmerges-gw6-zoo2-1 Started Stderr: Container roottestsystemmerges-gw6-zoo2-1 Started Stderr: Container roottestsystemmerges-gw6-zoo1-1 Started Stderr: Container roottestsystemmerges-gw6-zoo1-1 Started Stderr:time="2025-04-02T04:14:21Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:21Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:21Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:21Z" level=debug msg="otel error" error="" Wait ZooKeeper to start Wait ZooKeeper to start get_instance_ip instance_name=zoo1 get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-zoo1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.4.2, port:2181, use_ssl:False get_kazoo_client: zoo1, ip:172.16.4.2, port:2181, use_ssl:False Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT currentUser() on node Executing query SELECT currentUser() on node [gw8] PASSED test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottests3zerocopyreplication-gw0_default Creating Stderr: Network roottests3zerocopyreplication-gw0_default Creating Stderr: Network roottests3zerocopyreplication-gw0_default Created Stderr: Network roottests3zerocopyreplication-gw0_default Created Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Created Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Created Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Created Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Created Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Created Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Created Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Started Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Started Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Started Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Started Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Started Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Started Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Wait ZooKeeper to start Wait ZooKeeper to start get_instance_ip instance_name=zoo1 get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-zoo1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.6.4, port:2181, use_ssl:False get_kazoo_client: zoo1, ip:172.16.6.4, port:2181, use_ssl:False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestsystemstartstoplisten-gw9_default Creating Stderr: Network roottestsystemstartstoplisten-gw9_default Creating Stderr: Network roottestsystemstartstoplisten-gw9_default Created Stderr: Network roottestsystemstartstoplisten-gw9_default Created Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Started Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Started Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Started Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Started Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Started Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Started Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Wait ZooKeeper to start Wait ZooKeeper to start get_instance_ip instance_name=zoo1 get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-zoo1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.5.4, port:2181, use_ssl:False get_kazoo_client: zoo1, ip:172.16.5.4, port:2181, use_ssl:False Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT currentUser() on node Executing query SELECT currentUser() on node [gw8] PASSED test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network rootteststorageurl-gw4_default Creating Stderr: Network rootteststorageurl-gw4_default Creating Stderr: Network rootteststorageurl-gw4_default Created Stderr: Network rootteststorageurl-gw4_default Created Stderr: Container rootteststorageurl-gw4-nginx-1 Creating Stderr: Container rootteststorageurl-gw4-nginx-1 Creating Stderr: Container rootteststorageurl-gw4-nginx-1 Created Stderr: Container rootteststorageurl-gw4-nginx-1 Created Stderr: Container rootteststorageurl-gw4-nginx-1 Starting Stderr: Container rootteststorageurl-gw4-nginx-1 Starting Stderr: Container rootteststorageurl-gw4-nginx-1 Started Stderr: Container rootteststorageurl-gw4-nginx-1 Started Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml up -d --no-recreate] Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network rootteststorageredis-gw1_default Creating Stderr: Network rootteststorageredis-gw1_default Creating Stderr: Network rootteststorageredis-gw1_default Created Stderr: Network rootteststorageredis-gw1_default Created Stderr: Container rootteststorageredis-gw1-redis1-1 Creating Stderr: Container rootteststorageredis-gw1-redis1-1 Creating Stderr: Container rootteststorageredis-gw1-redis1-1 Created Stderr: Container rootteststorageredis-gw1-redis1-1 Created Stderr: Container rootteststorageredis-gw1-redis1-1 Starting Stderr: Container rootteststorageredis-gw1-redis1-1 Starting Stderr: Container rootteststorageredis-gw1-redis1-1 Started Stderr: Container rootteststorageredis-gw1-redis1-1 Started Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:22Z" level=debug msg="otel error" error="" [gw8] PASSED test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa Executing query SELECT currentUser() on node Executing query SELECT currentUser() on node Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused [gw8] PASSED test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key Executing query SELECT currentUser() on node Executing query SELECT currentUser() on node http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Command:[docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml stop --timeout 20] [gw8] PASSED test_ssh_keys_authentication/test.py::test_wrong_key Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Stderr: Container rootteststorageurl-gw4-nginx-1 Running Stderr: Container rootteststorageurl-gw4-nginx-1 Running Stderr: Container rootteststorageurl-gw4-node1-1 Creating Stderr: Container rootteststorageurl-gw4-node1-1 Creating Stderr: Container rootteststorageurl-gw4-node1-1 Created Stderr: Container rootteststorageurl-gw4-node1-1 Created Stderr: Container rootteststorageurl-gw4-node1-1 Starting Stderr: Container rootteststorageurl-gw4-node1-1 Starting Stderr: Container rootteststorageurl-gw4-node1-1 Started Stderr: Container rootteststorageurl-gw4-node1-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/rootteststorageurl-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurl-gw4-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/rootteststorageurl-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurl-gw4-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.7.3... Waiting for ClickHouse start in node1, ip: 172.16.7.3... http://localhost:None "GET /v1.46/containers/rootteststorageurl-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurl-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6ac527d0b523fce642dee886196451c2fd930e4bb77734785e3748fc63abbc3c/json HTTP/1.1" 200 None ClickHouse node1 started ClickHouse node1 started Executing query insert into table function url(url1) partition by column3 values (1, 2, 3), (3, 2, 1), (1, 3, 2) on node1 Executing query insert into table function url(url1) partition by column3 values (1, 2, 3), (3, 2, 1), (1, 3, 2) on node1 http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Executing query insert into table function url(url_file, url = 'http://nginx:80/CSV_file', format = 'CSV') values on node1 Executing query insert into table function url(url_file, url = 'http://nginx:80/CSV_file', format = 'CSV') values on node1 Executing query SELECT * FROM url('http://nginx:80/CSV_file', 'CSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 1 on node1 Executing query SELECT * FROM url('http://nginx:80/CSV_file', 'CSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 1 on node1 http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None Executing query SELECT * FROM url('http://nginx:80/CSV_file', 'CSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 4 on node1 Executing query SELECT * FROM url('http://nginx:80/CSV_file', 'CSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 4 on node1 Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9524a1e629df5406e504aba17472069e8d61f28d25dceb9c435d999e48095edc/json HTTP/1.1" 200 None ClickHouse ch1 started ClickHouse ch1 started Executing query CREATE DATABASE dictionaries ENGINE = Dictionary on ch1 Executing query CREATE DATABASE dictionaries ENGINE = Dictionary on ch1 Executing query CREATE TABLE dictionary_source (id UInt64, value UInt8) ENGINE = Memory on ch1 Executing query CREATE TABLE dictionary_source (id UInt64, value UInt8) ENGINE = Memory on ch1 Executing query SELECT * FROM url('http://nginx:80/CSV_file', 'CSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 16 on node1 Executing query SELECT * FROM url('http://nginx:80/CSV_file', 'CSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 16 on node1 run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo 127.0.0.1 localhost > /etc/hosts'] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo 127.0.0.1 localhost > /etc/hosts'] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo 127.0.0.1 localhost > /etc/hosts] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo 127.0.0.1 localhost > /etc/hosts] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo ::1 localhost >> /etc/hosts'] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo ::1 localhost >> /etc/hosts'] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo ::1 localhost >> /etc/hosts] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo ::1 localhost >> /etc/hosts] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo 127.255.255.255 lost_host >> /etc/hosts'] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo 127.255.255.255 lost_host >> /etc/hosts'] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo 127.255.255.255 lost_host >> /etc/hosts] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo 127.255.255.255 lost_host >> /etc/hosts] [gw4] PASSED test_storage_url/test.py::test_file_formats[CSV] test_storage_url/test.py::test_file_formats[JSONEachRow] Executing query insert into table function url(url_file, url = 'http://nginx:80/JSONEachRow_file', format = 'JSONEachRow') values on node1 Executing query insert into table function url(url_file, url = 'http://nginx:80/JSONEachRow_file', format = 'JSONEachRow') values on node1 Executing query SYSTEM DROP DNS CACHE on ch1 Executing query SYSTEM DROP DNS CACHE on ch1 Executing query SELECT * FROM url('http://nginx:80/JSONEachRow_file', 'JSONEachRow') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 1 on node1 Executing query SELECT * FROM url('http://nginx:80/JSONEachRow_file', 'JSONEachRow') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 1 on node1 Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT * FROM url('http://nginx:80/JSONEachRow_file', 'JSONEachRow') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 4 on node1 Executing query SELECT * FROM url('http://nginx:80/JSONEachRow_file', 'JSONEachRow') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 4 on node1 Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Connecting to 172.16.4.2(172.16.4.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Executing query SELECT * FROM url('http://nginx:80/JSONEachRow_file', 'JSONEachRow') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 16 on node1 Executing query SELECT * FROM url('http://nginx:80/JSONEachRow_file', 'JSONEachRow') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 16 on node1 Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.4.3, port:2181, use_ssl:False get_kazoo_client: zoo2, ip:172.16.4.3, port:2181, use_ssl:False Connecting to 172.16.4.3(172.16.4.3):2181, use_ssl: False Connecting to 172.16.4.3(172.16.4.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-zoo3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.4.4, port:2181, use_ssl:False get_kazoo_client: zoo3, ip:172.16.4.4, port:2181, use_ssl:False Connecting to 172.16.4.4(172.16.4.4):2181, use_ssl: False Connecting to 172.16.4.4(172.16.4.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost [gw4] PASSED test_storage_url/test.py::test_file_formats[JSONEachRow] test_storage_url/test.py::test_file_formats[Parquet] Executing query insert into table function url(url_file, url = 'http://nginx:80/Parquet_file', format = 'Parquet') values on node1 Executing query insert into table function url(url_file, url = 'http://nginx:80/Parquet_file', format = 'Parquet') values on node1 Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml up -d --no-recreate] Executing query SELECT * FROM url('http://nginx:80/Parquet_file', 'Parquet') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 1 on node1 Executing query SELECT * FROM url('http://nginx:80/Parquet_file', 'Parquet') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 1 on node1 Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT * FROM url('http://nginx:80/Parquet_file', 'Parquet') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 4 on node1 Executing query SELECT * FROM url('http://nginx:80/Parquet_file', 'Parquet') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 4 on node1 Executing query SELECT * FROM url('http://nginx:80/Parquet_file', 'Parquet') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 16 on node1 Executing query SELECT * FROM url('http://nginx:80/Parquet_file', 'Parquet') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 16 on node1 [gw4] PASSED test_storage_url/test.py::test_file_formats[Parquet] test_storage_url/test.py::test_file_formats[TSV] Executing query insert into table function url(url_file, url = 'http://nginx:80/TSV_file', format = 'TSV') values on node1 Executing query insert into table function url(url_file, url = 'http://nginx:80/TSV_file', format = 'TSV') values on node1 Stderr: mongo_no_cred Skipped - Image is already being pulled by mongo1 Stderr: mongo_no_cred Skipped - Image is already being pulled by mongo1 Stderr: mongo_secure Skipped - Image is already being pulled by mongo1 Stderr: mongo_secure Skipped - Image is already being pulled by mongo1 Stderr: mongo1 Pulling Stderr: mongo1 Pulling Stderr: node Pulling Stderr: node Pulling Stderr: 9cb31e2e37ea Already exists Stderr: 9cb31e2e37ea Already exists Stderr: 147a757512f6 Pulling fs layer Stderr: 147a757512f6 Pulling fs layer Stderr: 876012a5f4b7 Pulling fs layer Stderr: 876012a5f4b7 Pulling fs layer Stderr: 15b91eb705c7 Pulling fs layer Stderr: 15b91eb705c7 Pulling fs layer Stderr: caa49e3e29ad Pulling fs layer Stderr: caa49e3e29ad Pulling fs layer Stderr: 152a4f6c3783 Pulling fs layer Stderr: 152a4f6c3783 Pulling fs layer Stderr: 61401eeca3a3 Pulling fs layer Stderr: 61401eeca3a3 Pulling fs layer Stderr: 9f1c81570b11 Pulling fs layer Stderr: 9f1c81570b11 Pulling fs layer Stderr: 61401eeca3a3 Waiting Stderr: 61401eeca3a3 Waiting Stderr: caa49e3e29ad Waiting Stderr: caa49e3e29ad Waiting Stderr: 152a4f6c3783 Waiting Stderr: 152a4f6c3783 Waiting Stderr: 9f1c81570b11 Waiting Stderr: 9f1c81570b11 Waiting Stderr: 147a757512f6 Downloading [==================================================>] 1.782kB/1.782kB Stderr: 147a757512f6 Downloading [==================================================>] 1.782kB/1.782kB Stderr: 147a757512f6 Verifying Checksum Stderr: 147a757512f6 Verifying Checksum Stderr: 147a757512f6 Download complete Stderr: 147a757512f6 Download complete Stderr: 147a757512f6 Extracting [==================================================>] 1.782kB/1.782kB Stderr: 147a757512f6 Extracting [==================================================>] 1.782kB/1.782kB Stderr: 147a757512f6 Extracting [==================================================>] 1.782kB/1.782kB Stderr: 147a757512f6 Extracting [==================================================>] 1.782kB/1.782kB Stderr: 876012a5f4b7 Downloading [> ] 22.47kB/1.513MB Stderr: 876012a5f4b7 Downloading [> ] 22.47kB/1.513MB Stderr: 15b91eb705c7 Downloading [> ] 13.78kB/1.095MB Stderr: 15b91eb705c7 Downloading [> ] 13.78kB/1.095MB Stderr: 15b91eb705c7 Verifying Checksum Stderr: 15b91eb705c7 Verifying Checksum Stderr: 15b91eb705c7 Download complete Stderr: 15b91eb705c7 Download complete Stderr: 876012a5f4b7 Download complete Stderr: 876012a5f4b7 Download complete Stderr: caa49e3e29ad Downloading [==================================================>] 116B/116B Stderr: caa49e3e29ad Downloading [==================================================>] 116B/116B Stderr: caa49e3e29ad Download complete Stderr: caa49e3e29ad Download complete Stderr: 152a4f6c3783 Downloading [==================================================>] 263B/263B Stderr: 152a4f6c3783 Downloading [==================================================>] 263B/263B Stderr: 152a4f6c3783 Verifying Checksum Stderr: 152a4f6c3783 Verifying Checksum Stderr: 152a4f6c3783 Download complete Stderr: 152a4f6c3783 Download complete Stderr: 9f1c81570b11 Downloading [====================================> ] 3.647kB/4.997kB Stderr: 9f1c81570b11 Downloading [====================================> ] 3.647kB/4.997kB Stderr: 9f1c81570b11 Downloading [==================================================>] 4.997kB/4.997kB Stderr: 9f1c81570b11 Downloading [==================================================>] 4.997kB/4.997kB Stderr: 9f1c81570b11 Download complete Stderr: 9f1c81570b11 Download complete Stderr: 61401eeca3a3 Downloading [> ] 531.1kB/225.9MB Stderr: 61401eeca3a3 Downloading [> ] 531.1kB/225.9MB Stderr: 147a757512f6 Pull complete Stderr: 147a757512f6 Pull complete Stderr: 876012a5f4b7 Extracting [=> ] 32.77kB/1.513MB Stderr: 876012a5f4b7 Extracting [=> ] 32.77kB/1.513MB Stderr: 61401eeca3a3 Downloading [========> ] 40.06MB/225.9MB Stderr: 61401eeca3a3 Downloading [========> ] 40.06MB/225.9MB Stderr: 876012a5f4b7 Extracting [==========================================> ] 1.278MB/1.513MB Stderr: 876012a5f4b7 Extracting [==========================================> ] 1.278MB/1.513MB Stderr: 876012a5f4b7 Extracting [==================================================>] 1.513MB/1.513MB Stderr: 876012a5f4b7 Extracting [==================================================>] 1.513MB/1.513MB Stderr: 876012a5f4b7 Pull complete Stderr: 876012a5f4b7 Pull complete Stderr: 15b91eb705c7 Extracting [=> ] 32.77kB/1.095MB Stderr: 15b91eb705c7 Extracting [=> ] 32.77kB/1.095MB Stderr: node Pulled Stderr: node Pulled Stderr: 15b91eb705c7 Extracting [==================================================>] 1.095MB/1.095MB Stderr: 15b91eb705c7 Extracting [==================================================>] 1.095MB/1.095MB Stderr: 15b91eb705c7 Extracting [==================================================>] 1.095MB/1.095MB Stderr: 15b91eb705c7 Extracting [==================================================>] 1.095MB/1.095MB Stderr: 61401eeca3a3 Downloading [=================> ] 80.17MB/225.9MB Stderr: 61401eeca3a3 Downloading [=================> ] 80.17MB/225.9MB Stderr: 15b91eb705c7 Pull complete Stderr: 15b91eb705c7 Pull complete Stderr: caa49e3e29ad Extracting [==================================================>] 116B/116B Stderr: caa49e3e29ad Extracting [==================================================>] 116B/116B Stderr: caa49e3e29ad Extracting [==================================================>] 116B/116B Stderr: caa49e3e29ad Extracting [==================================================>] 116B/116B Stderr: caa49e3e29ad Pull complete Stderr: caa49e3e29ad Pull complete Stderr: 152a4f6c3783 Extracting [==================================================>] 263B/263B Stderr: 152a4f6c3783 Extracting [==================================================>] 263B/263B Stderr: 152a4f6c3783 Extracting [==================================================>] 263B/263B Stderr: 152a4f6c3783 Extracting [==================================================>] 263B/263B Stderr: 152a4f6c3783 Pull complete Stderr: 152a4f6c3783 Pull complete Stderr: 61401eeca3a3 Downloading [=========================> ] 116.4MB/225.9MB Stderr: 61401eeca3a3 Downloading [=========================> ] 116.4MB/225.9MB Stderr: 61401eeca3a3 Downloading [===================================> ] 159.3MB/225.9MB Stderr: 61401eeca3a3 Downloading [===================================> ] 159.3MB/225.9MB Stderr: 61401eeca3a3 Downloading [===========================================> ] 194.4MB/225.9MB Stderr: 61401eeca3a3 Downloading [===========================================> ] 194.4MB/225.9MB Stderr: 61401eeca3a3 Download complete Stderr: 61401eeca3a3 Download complete Stderr: 61401eeca3a3 Extracting [> ] 557.1kB/225.9MB Stderr: 61401eeca3a3 Extracting [> ] 557.1kB/225.9MB Stderr: 61401eeca3a3 Extracting [=> ] 7.799MB/225.9MB Stderr: 61401eeca3a3 Extracting [=> ] 7.799MB/225.9MB Stderr: 61401eeca3a3 Extracting [==> ] 12.81MB/225.9MB Stderr: 61401eeca3a3 Extracting [==> ] 12.81MB/225.9MB Stderr: 61401eeca3a3 Extracting [====> ] 19.5MB/225.9MB Stderr: 61401eeca3a3 Extracting [====> ] 19.5MB/225.9MB Stderr: 61401eeca3a3 Extracting [=====> ] 26.18MB/225.9MB Stderr: 61401eeca3a3 Extracting [=====> ] 26.18MB/225.9MB Stderr: 61401eeca3a3 Extracting [=======> ] 35.09MB/225.9MB Stderr: 61401eeca3a3 Extracting [=======> ] 35.09MB/225.9MB Stderr: 61401eeca3a3 Extracting [==========> ] 45.68MB/225.9MB Stderr: 61401eeca3a3 Extracting [==========> ] 45.68MB/225.9MB Stderr: 61401eeca3a3 Extracting [==========> ] 49.02MB/225.9MB Stderr: 61401eeca3a3 Extracting [==========> ] 49.02MB/225.9MB Stderr: 61401eeca3a3 Extracting [============> ] 58.49MB/225.9MB Stderr: 61401eeca3a3 Extracting [============> ] 58.49MB/225.9MB Stderr: 61401eeca3a3 Extracting [==============> ] 64.06MB/225.9MB Stderr: 61401eeca3a3 Extracting [==============> ] 64.06MB/225.9MB Stderr: 61401eeca3a3 Extracting [===============> ] 70.75MB/225.9MB Stderr: 61401eeca3a3 Extracting [===============> ] 70.75MB/225.9MB Stderr: 61401eeca3a3 Extracting [==================> ] 82.44MB/225.9MB Stderr: 61401eeca3a3 Extracting [==================> ] 82.44MB/225.9MB Stderr: 61401eeca3a3 Extracting [====================> ] 94.14MB/225.9MB Stderr: 61401eeca3a3 Extracting [====================> ] 94.14MB/225.9MB Stderr: 61401eeca3a3 Extracting [=====================> ] 98.6MB/225.9MB Stderr: 61401eeca3a3 Extracting [=====================> ] 98.6MB/225.9MB Stderr: 61401eeca3a3 Extracting [=======================> ] 107.5MB/225.9MB Stderr: 61401eeca3a3 Extracting [=======================> ] 107.5MB/225.9MB Stderr: 61401eeca3a3 Extracting [=========================> ] 115.3MB/225.9MB Stderr: 61401eeca3a3 Extracting [=========================> ] 115.3MB/225.9MB Stderr: 61401eeca3a3 Extracting [===========================> ] 124.2MB/225.9MB Stderr: 61401eeca3a3 Extracting [===========================> ] 124.2MB/225.9MB Stderr: 61401eeca3a3 Extracting [=============================> ] 132MB/225.9MB Stderr: 61401eeca3a3 Extracting [=============================> ] 132MB/225.9MB Stderr: 61401eeca3a3 Extracting [===============================> ] 141.5MB/225.9MB Stderr: 61401eeca3a3 Extracting [===============================> ] 141.5MB/225.9MB Stderr: 61401eeca3a3 Extracting [================================> ] 148.7MB/225.9MB Stderr: 61401eeca3a3 Extracting [================================> ] 148.7MB/225.9MB Stderr: 61401eeca3a3 Extracting [===================================> ] 158.2MB/225.9MB Stderr: 61401eeca3a3 Extracting [===================================> ] 158.2MB/225.9MB Stderr: 61401eeca3a3 Extracting [=====================================> ] 168.2MB/225.9MB Stderr: 61401eeca3a3 Extracting [=====================================> ] 168.2MB/225.9MB Stderr: 61401eeca3a3 Extracting [======================================> ] 175.5MB/225.9MB Stderr: 61401eeca3a3 Extracting [======================================> ] 175.5MB/225.9MB Stderr: 61401eeca3a3 Extracting [=========================================> ] 188.3MB/225.9MB Stderr: 61401eeca3a3 Extracting [=========================================> ] 188.3MB/225.9MB Stderr: 61401eeca3a3 Extracting [===========================================> ] 198.3MB/225.9MB Stderr: 61401eeca3a3 Extracting [===========================================> ] 198.3MB/225.9MB Stderr: 61401eeca3a3 Extracting [=============================================> ] 207.8MB/225.9MB Stderr: 61401eeca3a3 Extracting [=============================================> ] 207.8MB/225.9MB Stderr: 61401eeca3a3 Extracting [===============================================> ] 216.1MB/225.9MB Stderr: 61401eeca3a3 Extracting [===============================================> ] 216.1MB/225.9MB Stderr: 61401eeca3a3 Extracting [=================================================> ] 221.7MB/225.9MB Stderr: 61401eeca3a3 Extracting [=================================================> ] 221.7MB/225.9MB Stderr: 61401eeca3a3 Extracting [=================================================> ] 223.4MB/225.9MB Stderr: 61401eeca3a3 Extracting [=================================================> ] 223.4MB/225.9MB Stderr: 61401eeca3a3 Extracting [=================================================> ] 225.1MB/225.9MB Stderr: 61401eeca3a3 Extracting [=================================================> ] 225.1MB/225.9MB Stderr: 61401eeca3a3 Extracting [=================================================> ] 225.6MB/225.9MB Stderr: 61401eeca3a3 Extracting [=================================================> ] 225.6MB/225.9MB Stderr: 61401eeca3a3 Extracting [==================================================>] 225.9MB/225.9MB Stderr: 61401eeca3a3 Extracting [==================================================>] 225.9MB/225.9MB Stderr: 61401eeca3a3 Pull complete Stderr: 61401eeca3a3 Pull complete Stderr: 9f1c81570b11 Extracting [==================================================>] 4.997kB/4.997kB Stderr: 9f1c81570b11 Extracting [==================================================>] 4.997kB/4.997kB Stderr: 9f1c81570b11 Extracting [==================================================>] 4.997kB/4.997kB Stderr: 9f1c81570b11 Extracting [==================================================>] 4.997kB/4.997kB Stderr: 9f1c81570b11 Pull complete Stderr: 9f1c81570b11 Pull complete Stderr: mongo1 Pulled Stderr: mongo1 Pulled Setup Mongo Setup Mongo Command:[docker compose --project-name roottesttablefunctionmongodblegacy-gw3 --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml --verbose up -d] Command:[docker compose --project-name roottesttablefunctionmongodblegacy-gw3 --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml --verbose up -d] Executing query SELECT * FROM url('http://nginx:80/TSV_file', 'TSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 1 on node1 Executing query SELECT * FROM url('http://nginx:80/TSV_file', 'TSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 1 on node1 Executing query SELECT * FROM url('http://nginx:80/TSV_file', 'TSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 4 on node1 Executing query SELECT * FROM url('http://nginx:80/TSV_file', 'TSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 4 on node1 Stderr: Container roottestsystemmerges-gw6-zoo3-1 Running Stderr: Container roottestsystemmerges-gw6-zoo3-1 Running Stderr: Container roottestsystemmerges-gw6-zoo2-1 Running Stderr: Container roottestsystemmerges-gw6-zoo2-1 Running Stderr: Container roottestsystemmerges-gw6-zoo1-1 Running Stderr: Container roottestsystemmerges-gw6-zoo1-1 Running Stderr: Container roottestsystemmerges-gw6-node1-1 Creating Stderr: Container roottestsystemmerges-gw6-node1-1 Creating Stderr: Container roottestsystemmerges-gw6-node2-1 Creating Stderr: Container roottestsystemmerges-gw6-node2-1 Creating Stderr: Container roottestsystemmerges-gw6-node2-1 Created Stderr: Container roottestsystemmerges-gw6-node2-1 Created Stderr: Container roottestsystemmerges-gw6-node1-1 Created Stderr: Container roottestsystemmerges-gw6-node1-1 Created Stderr: Container roottestsystemmerges-gw6-node1-1 Starting Stderr: Container roottestsystemmerges-gw6-node1-1 Starting Stderr: Container roottestsystemmerges-gw6-node2-1 Starting Stderr: Container roottestsystemmerges-gw6-node2-1 Starting Stderr: Container roottestsystemmerges-gw6-node1-1 Started Stderr: Container roottestsystemmerges-gw6-node1-1 Started Stderr: Container roottestsystemmerges-gw6-node2-1 Started Stderr: Container roottestsystemmerges-gw6-node2-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.4.5... Waiting for ClickHouse start in node1, ip: 172.16.4.5... http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None Executing query SELECT * FROM url('http://nginx:80/TSV_file', 'TSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 16 on node1 Executing query SELECT * FROM url('http://nginx:80/TSV_file', 'TSV') LIMIT 10 SETTINGS remote_read_min_bytes_for_seek = 1, max_read_buffer_size = 1, max_download_buffer_size = 1, max_download_threads = 16 on node1 http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None [gw4] PASSED test_storage_url/test.py::test_file_formats[TSV] test_storage_url/test.py::test_partition_by Executing query select * from url('http://nginx:80/test_1', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query select * from url('http://nginx:80/test_1', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Executing query select * from url('http://nginx:80/test_2', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query select * from url('http://nginx:80/test_2', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.6.3, port:2181, use_ssl:False get_kazoo_client: zoo2, ip:172.16.6.3, port:2181, use_ssl:False Connecting to 172.16.6.3(172.16.6.3):2181, use_ssl: False Connecting to 172.16.6.3(172.16.6.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-zoo3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.6.2, port:2181, use_ssl:False get_kazoo_client: zoo3, ip:172.16.6.2, port:2181, use_ssl:False Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Executing query select * from url('http://nginx:80/test_3', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query select * from url('http://nginx:80/test_3', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Stderr:time="2025-04-02T04:14:24Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:24Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Creating Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Creating Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Created Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Started Stderr:time="2025-04-02T04:14:25Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:25Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:25Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:25Z" level=debug msg="otel error" error="" Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') Trying to create Minio instance by command docker compose --project-name roottests3zerocopyreplication-gw0 --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Trying to create Minio instance by command docker compose --project-name roottests3zerocopyreplication-gw0 --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Command:[docker compose --project-name roottests3zerocopyreplication-gw0 --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Command:[docker compose --project-name roottests3zerocopyreplication-gw0 --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb0f8baaf6acd406ff979bee1dc95680efd2a238a3731eb022b0dd5d6eb0ca67/json HTTP/1.1" 200 None ClickHouse node1 started ClickHouse node1 started get_instance_ip instance_name=node2 get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.4.6... Waiting for ClickHouse start in node2, ip: 172.16.4.6... http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemmerges-gw6-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc0f2698175f2ef4c2be2b2b0426ecfb0e15f61b6e32da4b2e66a0e27d5e11d0/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc0f2698175f2ef4c2be2b2b0426ecfb0e15f61b6e32da4b2e66a0e27d5e11d0/json HTTP/1.1" 200 None ClickHouse node2 started ClickHouse node2 started Executing query CREATE DATABASE test ENGINE=Ordinary on node1 Executing query CREATE DATABASE test ENGINE=Ordinary on node1 Executing query CREATE USER OR REPLACE u1 on node1 [gw4] PASSED test_storage_url/test.py::test_partition_by Executing query CREATE USER OR REPLACE u1 on node1 test_storage_url/test.py::test_table_function_url_access_rights Executing query CREATE DATABASE test ENGINE=Ordinary on node2 Executing query CREATE DATABASE test ENGINE=Ordinary on node2 Executing query SELECT * FROM remote('lost_host', 'system', 'one') on ch1 Executing query SELECT * FROM remote('lost_host', 'system', 'one') on ch1 Executing query SELECT * FROM url('http://nginx:80/test_1', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query SELECT * FROM url('http://nginx:80/test_1', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query create table test.merge_simple (a Int64) engine=MergeTree() order by tuple() on node1 Executing query create table test.merge_simple (a Int64) engine=MergeTree() order by tuple() on node1 Executing query CREATE TABLE distributed_lost_host (dummy UInt8) ENGINE = Distributed(lost_host_cluster, 'system', 'one') on ch1 Executing query CREATE TABLE distributed_lost_host (dummy UInt8) ENGINE = Distributed(lost_host_cluster, 'system', 'one') on ch1 Executing query SELECT * FROM url('http://nginx:80/test_1', 'TSV') on node1 Executing query SELECT * FROM url('http://nginx:80/test_1', 'TSV') on node1 Executing query INSERT INTO test.merge_simple VALUES (1) on node1 Executing query INSERT INTO test.merge_simple VALUES (1) on node1 Executing query SELECT * FROM distributed_lost_host on ch1 Executing query SELECT * FROM distributed_lost_host on ch1 Executing query DESCRIBE TABLE url('http://nginx:80/test_1', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query DESCRIBE TABLE url('http://nginx:80/test_1', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query INSERT INTO test.merge_simple VALUES (2) on node1 Executing query INSERT INTO test.merge_simple VALUES (2) on node1 run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo 127.0.0.1 localhost > /etc/hosts'] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo 127.0.0.1 localhost > /etc/hosts'] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo 127.0.0.1 localhost > /etc/hosts] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo 127.0.0.1 localhost > /etc/hosts] Executing query DESCRIBE TABLE url('http://nginx:80/not-exist', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query DESCRIBE TABLE url('http://nginx:80/not-exist', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo ::1 localhost >> /etc/hosts'] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo ::1 localhost >> /etc/hosts'] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo ::1 localhost >> /etc/hosts] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo ::1 localhost >> /etc/hosts] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo 127.0.0.1 lost_host >> /etc/hosts'] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo 127.0.0.1 lost_host >> /etc/hosts'] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo 127.0.0.1 lost_host >> /etc/hosts] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo 127.0.0.1 lost_host >> /etc/hosts] Executing query SYSTEM DROP DNS CACHE on ch1 Executing query SYSTEM DROP DNS CACHE on ch1 Executing query DESCRIBE TABLE url('http://nginx:80/test_1', 'TSV') on node1 Executing query DESCRIBE TABLE url('http://nginx:80/test_1', 'TSV') on node1 Executing query SELECT * FROM remote('lost_host', 'system', 'one') on ch1 Executing query SELECT * FROM remote('lost_host', 'system', 'one') on ch1 Executing query INSERT INTO test.merge_simple VALUES (3) on node1 Executing query INSERT INTO test.merge_simple VALUES (3) on node1 Executing query GRANT URL ON *.* TO u1 on node1 Executing query GRANT URL ON *.* TO u1 on node1 Executing query SELECT * FROM distributed_lost_host on ch1 Executing query SELECT * FROM distributed_lost_host on ch1 Executing query alter table test.merge_simple add column b int materialized sleepEachRow(3) on node1 Executing query alter table test.merge_simple add column b int materialized sleepEachRow(3) on node1 Executing query DESCRIBE TABLE url('http://nginx:80/test_1', 'TSV') on node1 Executing query DESCRIBE TABLE url('http://nginx:80/test_1', 'TSV') on node1 Executing query SELECT DISTINCT host_name, host_address FROM system.clusters WHERE cluster='lost_host_cluster' on ch1 Executing query SELECT DISTINCT host_name, host_address FROM system.clusters WHERE cluster='lost_host_cluster' on ch1 Executing query OPTIMIZE TABLE test.merge_simple on node1 Executing query OPTIMIZE TABLE test.merge_simple on node1 Executing query select count() from system.merges where table='merge_simple' on node1 Executing query select count() from system.merges where table='merge_simple' on node1 [gw7] PASSED test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "ro" > /etc/clickhouse-server/config.d/macros.xml'] run container_id:roottestsystemqueries-gw7-ch1-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "ro" > /etc/clickhouse-server/config.d/macros.xml'] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo "ro" > /etc/clickhouse-server/config.d/macros.xml] Command:[docker exec -u root --privileged roottestsystemqueries-gw7-ch1-1 bash -c echo "ro" > /etc/clickhouse-server/config.d/macros.xml] Stderr:time="2025-04-02T04:14:25Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:25Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Volume "roottests3zerocopyreplication-gw0_data1-1" Creating Stderr: Volume "roottests3zerocopyreplication-gw0_data1-1" Creating Stderr: Volume "roottests3zerocopyreplication-gw0_data1-1" Created Stderr: Volume "roottests3zerocopyreplication-gw0_data1-1" Created Stderr:time="2025-04-02T04:14:25Z" level=warning msg="Found orphan containers ([roottests3zerocopyreplication-gw0-zoo1-1 roottests3zerocopyreplication-gw0-zoo2-1 roottests3zerocopyreplication-gw0-zoo3-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." Stderr:time="2025-04-02T04:14:25Z" level=warning msg="Found orphan containers ([roottests3zerocopyreplication-gw0-zoo1-1 roottests3zerocopyreplication-gw0-zoo2-1 roottests3zerocopyreplication-gw0-zoo3-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Created Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Created Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Created Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Created Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Created Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Created Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Created Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Created Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Started Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Started Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Started Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Started Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Started Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Started Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Started Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Started [gw4] PASSED test_storage_url/test.py::test_table_function_url_access_rights Stderr:time="2025-04-02T04:14:25Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:25Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:25Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:25Z" level=debug msg="otel error" error="" Trying to connect to Minio... Trying to connect to Minio... get_instance_ip instance_name=minio1 get_instance_ip instance_name=minio1 test_storage_url/test.py::test_url_cluster Executing query select * from urlCluster('test_cluster_two_shards', 'http://nginx:80/test_1', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query select * from urlCluster('test_cluster_two_shards', 'http://nginx:80/test_1', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-minio1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-minio1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=proxy1 get_instance_ip instance_name=proxy1 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-proxy1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-proxy1-1/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.6.7:9001 Starting new HTTP connection (1): 172.16.6.7:9001 Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (2): 172.16.6.7:9001 Starting new HTTP connection (2): 172.16.6.7:9001 Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (3): 172.16.6.7:9001 Starting new HTTP connection (3): 172.16.6.7:9001 Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (4): 172.16.6.7:9001 Starting new HTTP connection (4): 172.16.6.7:9001 Can't connect to Minio: HTTPConnectionPool(host='172.16.6.7', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Can't connect to Minio: HTTPConnectionPool(host='172.16.6.7', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Executing query SYSTEM RELOAD CONFIG on ch1 Executing query SYSTEM RELOAD CONFIG on ch1 Executing query SELECT database, table, num_parts, source_part_names, source_part_paths, result_part_name, result_part_path, partition_id, is_mutation FROM system.merges WHERE table = 'merge_simple' on node1 Executing query SELECT database, table, num_parts, source_part_names, source_part_paths, result_part_name, result_part_path, partition_id, is_mutation FROM system.merges WHERE table = 'merge_simple' on node1 Executing query select * from urlCluster('test_cluster_two_shards', 'http://nginx:80/test_2', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query select * from urlCluster('test_cluster_two_shards', 'http://nginx:80/test_2', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query select * from system.macros on ch1 Executing query select * from system.macros on ch1 [gw7] PASSED test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY Executing query SYSTEM RELOAD DICTIONARIES on ch1 Executing query SYSTEM RELOAD DICTIONARIES on ch1 Executing query select * from urlCluster('test_cluster_two_shards', 'http://nginx:80/test_3', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query select * from urlCluster('test_cluster_two_shards', 'http://nginx:80/test_3', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32') on node1 Executing query SELECT dictHas('clickhouse_flat', toUInt64(0)), dictHas('clickhouse_flat', toUInt64(1)) on ch1 Executing query SELECT dictHas('clickhouse_flat', toUInt64(0)), dictHas('clickhouse_flat', toUInt64(1)) on ch1 [gw4] PASSED test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection Executing query select * from urlCluster(test_cluster_one_shard_three_replicas_localhost, test_url) on node1 Executing query select * from urlCluster(test_cluster_one_shard_three_replicas_localhost, test_url) on node1 Executing query select * from urlCluster(test_cluster_one_shard_three_replicas_localhost, test_url, structure='auto') on node1 Executing query select * from urlCluster(test_cluster_one_shard_three_replicas_localhost, test_url, structure='auto') on node1 Stderr: Container roottestsshkeysauthentication-gw8-node-1 Stopping Stderr: Container roottestsshkeysauthentication-gw8-node-1 Stopping Stderr: Container roottestsshkeysauthentication-gw8-node-1 Stopped Stderr: Container roottestsshkeysauthentication-gw8-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query INSERT INTO dictionary_source VALUES (0, 0) on ch1 Executing query INSERT INTO dictionary_source VALUES (0, 0) on ch1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/.env --project-name roottestsshkeysauthentication-gw8 --file /ClickHouse/tests/integration/test_ssh_keys_authentication/_instances-0-gw8/node/docker-compose.yml down --volumes] Executing query SELECT dictGetUInt8('clickhouse_cache', 'value', toUInt64(0)), dictHas('clickhouse_cache', toUInt64(1)) on ch1 Executing query SELECT dictGetUInt8('clickhouse_cache', 'value', toUInt64(0)), dictHas('clickhouse_cache', toUInt64(1)) on ch1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml stop --timeout 20] [gw4] PASSED test_storage_url/test.py::test_url_cluster_with_named_collection Executing query INSERT INTO dictionary_source VALUES (1, 1) on ch1 Executing query INSERT INTO dictionary_source VALUES (1, 1) on ch1 Executing query SELECT dictGetUInt8('clickhouse_cache', 'value', toUInt64(0)), dictHas('clickhouse_cache', toUInt64(1)) on ch1 Executing query SELECT dictGetUInt8('clickhouse_cache', 'value', toUInt64(0)), dictHas('clickhouse_cache', toUInt64(1)) on ch1 Executing query SYSTEM RELOAD DICTIONARY clickhouse_cache on ch1 Executing query SYSTEM RELOAD DICTIONARY clickhouse_cache on ch1 Executing query SELECT dictGetUInt8('clickhouse_cache', 'value', toUInt64(0)), dictGetUInt8('clickhouse_cache', 'value', toUInt64(1)) on ch1 Executing query SELECT dictGetUInt8('clickhouse_cache', 'value', toUInt64(0)), dictGetUInt8('clickhouse_cache', 'value', toUInt64(1)) on ch1 Executing query SELECT dictHas('clickhouse_flat', toUInt64(0)), dictHas('clickhouse_flat', toUInt64(1)) on ch1 Executing query SELECT dictHas('clickhouse_flat', toUInt64(0)), dictHas('clickhouse_flat', toUInt64(1)) on ch1 Executing query SYSTEM RELOAD DICTIONARIES on ch1 Executing query SYSTEM RELOAD DICTIONARIES on ch1 Starting new HTTP connection (5): 172.16.6.7:9001 Starting new HTTP connection (5): 172.16.6.7:9001 http://172.16.6.7:9001 "GET / HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET / HTTP/1.1" 200 0 Connected to Minio. Connected to Minio. http://172.16.6.7:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.6.7:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.6.7:9001 "PUT /root HTTP/1.1" 200 0 http://172.16.6.7:9001 "PUT /root HTTP/1.1" 200 0 S3 bucket 'root' created S3 bucket 'root' created http://172.16.6.7:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.6.7:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.6.7:9001 "PUT /root2 HTTP/1.1" 200 0 http://172.16.6.7:9001 "PUT /root2 HTTP/1.1" 200 0 S3 bucket 'root2' created S3 bucket 'root2' created ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml up -d --no-recreate] Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Stderr: Container roottestsshkeysauthentication-gw8-node-1 Stopping Stderr: Container roottestsshkeysauthentication-gw8-node-1 Stopping Stderr: Container roottestsshkeysauthentication-gw8-node-1 Stopped Stderr: Container roottestsshkeysauthentication-gw8-node-1 Stopped Stderr: Container roottestsshkeysauthentication-gw8-node-1 Removing Stderr: Container roottestsshkeysauthentication-gw8-node-1 Removing Stderr: Container roottestsshkeysauthentication-gw8-node-1 Removed Stderr: Container roottestsshkeysauthentication-gw8-node-1 Removed Stderr: Network roottestsshkeysauthentication-gw8_default Removing Stderr: Network roottestsshkeysauthentication-gw8_default Removing Stderr: Network roottestsshkeysauthentication-gw8_default Removed Stderr: Network roottestsshkeysauthentication-gw8_default Removed Cleanup called Cleanup called Docker networks for project roottestsshkeysauthentication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsshkeysauthentication-gw8 are NETWORK ID NAME DRIVER SCOPE Executing query SELECT dictGetUInt8('clickhouse_cache', 'value', toUInt64(0)), dictGetUInt8('clickhouse_cache', 'value', toUInt64(1)) on ch1 Executing query SELECT dictGetUInt8('clickhouse_cache', 'value', toUInt64(0)), dictGetUInt8('clickhouse_cache', 'value', toUInt64(1)) on ch1 Docker containers for project roottestsshkeysauthentication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsshkeysauthentication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsshkeysauthentication-gw8 are DRIVER VOLUME NAME Docker volumes for project roottestsshkeysauthentication-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsshkeysauthentication-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsshkeysauthentication-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.3.2, port:2181, use_ssl:False get_kazoo_client: zoo2, ip:172.16.3.2, port:2181, use_ssl:False Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsshkeysauthentication-gw8 No running containers for project: roottestsshkeysauthentication-gw8 Trying to prune unused networks... Trying to prune unused networks... Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:10 Stdout:10 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query SELECT dictGetUInt8('clickhouse_flat', 'value', toUInt64(0)), dictGetUInt8('clickhouse_flat', 'value', toUInt64(1)) on ch1 Executing query SELECT dictGetUInt8('clickhouse_flat', 'value', toUInt64(0)), dictGetUInt8('clickhouse_flat', 'value', toUInt64(1)) on ch1 Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 10 Volumes pruned: 10 test_storage_url_http_headers/test.py::test_storage_url_http_headers Running tests in /ClickHouse/tests/integration/test_storage_url_http_headers/test.py Running tests in /ClickHouse/tests/integration/test_storage_url_http_headers/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project rootteststorageurlhttpheaders-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurlhttpheaders-gw8 are NETWORK ID NAME DRIVER SCOPE Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-zoo3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.3.3, port:2181, use_ssl:False get_kazoo_client: zoo3, ip:172.16.3.3, port:2181, use_ssl:False Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Docker containers for project rootteststorageurlhttpheaders-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurlhttpheaders-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-zoo2-1/json HTTP/1.1" 200 None Docker volumes for project rootteststorageurlhttpheaders-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurlhttpheaders-gw8 are DRIVER VOLUME NAME Cleanup called Cleanup called get_kazoo_client: zoo2, ip:172.16.5.2, port:2181, use_ssl:False get_kazoo_client: zoo2, ip:172.16.5.2, port:2181, use_ssl:False Connecting to 172.16.5.2(172.16.5.2):2181, use_ssl: False Connecting to 172.16.5.2(172.16.5.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost [gw7] PASSED test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 test_system_queries/test.py::test_system_flush_logs Docker networks for project rootteststorageurlhttpheaders-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurlhttpheaders-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststorageurlhttpheaders-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurlhttpheaders-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststorageurlhttpheaders-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurlhttpheaders-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageurlhttpheaders-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageurlhttpheaders-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageurlhttpheaders-gw8 No running containers for project: rootteststorageurlhttpheaders-gw8 Trying to prune unused networks... Trying to prune unused networks... Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-zoo3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.5.3, port:2181, use_ssl:False get_kazoo_client: zoo3, ip:172.16.5.3, port:2181, use_ssl:False Connecting to 172.16.5.3(172.16.5.3):2181, use_ssl: False Connecting to 172.16.5.3(172.16.5.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Sending request(xid=2): Close() Sending request(xid=2): Close() Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') Trying to create Minio instance by command docker compose --project-name roottests3cluster-gw5 --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Trying to create Minio instance by command docker compose --project-name roottests3cluster-gw5 --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Command:[docker compose --project-name roottests3cluster-gw5 --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Command:[docker compose --project-name roottests3cluster-gw5 --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:10 Stdout:10 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 10 Volumes pruned: 10 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/configs/config.d Copy custom test config files [] to /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/database Setup database dir /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/database Setup logs dir /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/logs Setup logs dir /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml pull] Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml up -d --no-recreate] Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Running Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Running Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Running Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Running Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Running Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Running Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Running Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Running Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Running Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Running Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Running Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Running Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Running Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Running Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Creating Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Created Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Created Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Created Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Created Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Starting Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Started Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Started Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Started Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.6.9... Waiting for ClickHouse start in node1, ip: 172.16.6.9... http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Running Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Running Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Running Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Running Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Running Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Running Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Creating Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Created Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Starting Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Started Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Started Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Started Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=main_node get_instance_ip instance_name=main_node http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-main_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-main_node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=main_node get_instance_ip instance_name=main_node http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-main_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-main_node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in main_node, ip: 172.16.5.5... Waiting for ClickHouse start in main_node, ip: 172.16.5.5... http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-main_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-main_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d9a82580d21ba1eada574e6d7abe4a4c5b7f4bfd66b0ad0af3d09e1ac7fbfb80/json HTTP/1.1" 200 None ClickHouse node1 started ClickHouse node1 started get_instance_ip instance_name=node2 get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.6.10... Waiting for ClickHouse start in node2, ip: 172.16.6.10... http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3zerocopyreplication-gw0-node2-1/json HTTP/1.1" 200 None Executing query SELECT 1 FORMAT Null; SET log_queries = 0; SYSTEM FLUSH LOGS; SELECT count() FROM system.query_log; on ch1 Executing query SELECT 1 FORMAT Null; SET log_queries = 0; SYSTEM FLUSH LOGS; SELECT count() FROM system.query_log; on ch1 http://localhost:None "GET /v1.46/containers/71c36ee5cd900ea4e716b078908769bfbcde41f1a90005f86365e99543490d9e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/71c36ee5cd900ea4e716b078908769bfbcde41f1a90005f86365e99543490d9e/json HTTP/1.1" 200 None ClickHouse node2 started ClickHouse node2 started Cluster started Cluster started Executing query CREATE TABLE s3_test ON CLUSTER test_cluster (id UInt32, value String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/s3_test', '{replica}') ORDER BY id SETTINGS storage_policy='s3' on node1 Executing query CREATE TABLE s3_test ON CLUSTER test_cluster (id UInt32, value String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/s3_test', '{replica}') ORDER BY id SETTINGS storage_policy='s3' on node1 http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None Stderr:time="2025-04-02T04:14:27Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:27Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Volume "roottests3cluster-gw5_data1-1" Creating Stderr: Volume "roottests3cluster-gw5_data1-1" Creating Stderr: Volume "roottests3cluster-gw5_data1-1" Created Stderr: Volume "roottests3cluster-gw5_data1-1" Created Stderr:time="2025-04-02T04:14:27Z" level=warning msg="Found orphan containers ([roottests3cluster-gw5-zoo3-1 roottests3cluster-gw5-zoo1-1 roottests3cluster-gw5-zoo2-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." Stderr:time="2025-04-02T04:14:27Z" level=warning msg="Found orphan containers ([roottests3cluster-gw5-zoo3-1 roottests3cluster-gw5-zoo1-1 roottests3cluster-gw5-zoo2-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." Stderr: Container roottests3cluster-gw5-proxy1-1 Creating Stderr: Container roottests3cluster-gw5-proxy1-1 Creating Stderr: Container roottests3cluster-gw5-proxy2-1 Creating Stderr: Container roottests3cluster-gw5-proxy2-1 Creating Stderr: Container roottests3cluster-gw5-proxy1-1 Created Stderr: Container roottests3cluster-gw5-proxy1-1 Created Stderr: Container roottests3cluster-gw5-proxy2-1 Created Stderr: Container roottests3cluster-gw5-proxy2-1 Created Stderr: Container roottests3cluster-gw5-minio1-1 Creating Stderr: Container roottests3cluster-gw5-minio1-1 Creating Stderr: Container roottests3cluster-gw5-resolver-1 Creating Stderr: Container roottests3cluster-gw5-resolver-1 Creating Stderr: Container roottests3cluster-gw5-resolver-1 Created Stderr: Container roottests3cluster-gw5-resolver-1 Created Stderr: Container roottests3cluster-gw5-minio1-1 Created Stderr: Container roottests3cluster-gw5-minio1-1 Created Stderr: Container roottests3cluster-gw5-proxy2-1 Starting Stderr: Container roottests3cluster-gw5-proxy2-1 Starting Stderr: Container roottests3cluster-gw5-proxy1-1 Starting Stderr: Container roottests3cluster-gw5-proxy1-1 Starting Stderr: Container roottests3cluster-gw5-proxy2-1 Started Stderr: Container roottests3cluster-gw5-proxy2-1 Started Stderr: Container roottests3cluster-gw5-proxy1-1 Started Stderr: Container roottests3cluster-gw5-proxy1-1 Started Stderr: Container roottests3cluster-gw5-minio1-1 Starting Stderr: Container roottests3cluster-gw5-minio1-1 Starting Stderr: Container roottests3cluster-gw5-resolver-1 Starting Stderr: Container roottests3cluster-gw5-resolver-1 Starting Stderr: Container roottests3cluster-gw5-resolver-1 Started Stderr: Container roottests3cluster-gw5-resolver-1 Started Stderr: Container roottests3cluster-gw5-minio1-1 Started Stderr: Container roottests3cluster-gw5-minio1-1 Started Stderr:time="2025-04-02T04:14:28Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:28Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:28Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:28Z" level=debug msg="otel error" error="" Trying to connect to Minio... Trying to connect to Minio... get_instance_ip instance_name=minio1 get_instance_ip instance_name=minio1 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-minio1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-minio1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=proxy1 get_instance_ip instance_name=proxy1 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-proxy1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-proxy1-1/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.3.7:9001 Starting new HTTP connection (1): 172.16.3.7:9001 Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (2): 172.16.3.7:9001 Starting new HTTP connection (2): 172.16.3.7:9001 Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (3): 172.16.3.7:9001 Starting new HTTP connection (3): 172.16.3.7:9001 Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (4): 172.16.3.7:9001 Starting new HTTP connection (4): 172.16.3.7:9001 Can't connect to Minio: HTTPConnectionPool(host='172.16.3.7', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Can't connect to Minio: HTTPConnectionPool(host='172.16.3.7', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4a55411e446783c1f379af9c69fdc9cd3033c05883f99791f073121da99a7ced/json HTTP/1.1" 200 None ClickHouse main_node started ClickHouse main_node started get_instance_ip instance_name=backup_node get_instance_ip instance_name=backup_node http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-backup_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-backup_node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=backup_node get_instance_ip instance_name=backup_node http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-backup_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-backup_node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in backup_node, ip: 172.16.5.6... Waiting for ClickHouse start in backup_node, ip: 172.16.5.6... http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-backup_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemstartstoplisten-gw9-backup_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c333987b123b732ff95b67b079a3852fb173c9475c8ee3152e5d923c9e6d7af2/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c333987b123b732ff95b67b079a3852fb173c9475c8ee3152e5d923c9e6d7af2/json HTTP/1.1" 200 None ClickHouse backup_node started ClickHouse backup_node started Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query INSERT INTO s3_test VALUES (0,'data'),(1,'data') on node1 Executing query INSERT INTO s3_test VALUES (0,'data'),(1,'data') on node1 Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SYSTEM SYNC REPLICA s3_test on node2 Executing query SYSTEM SYNC REPLICA s3_test on node2 Executing query SELECT * FROM s3_test order by id FORMAT Values on node1 Executing query SELECT * FROM s3_test order by id FORMAT Values on node1 Executing query SELECT * FROM s3_test order by id FORMAT Values on node2 Executing query SELECT * FROM s3_test order by id FORMAT Values on node2 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN QUERIES ALL on main_node Executing query SYSTEM STOP LISTEN QUERIES ALL on main_node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query INSERT INTO s3_test VALUES (2,'data'),(3,'data') on node2 Executing query INSERT INTO s3_test VALUES (2,'data'),(3,'data') on node2 Executing query SYSTEM SYNC REPLICA s3_test on node1 Executing query SYSTEM SYNC REPLICA s3_test on node1 Executing query SELECT * FROM s3_test order by id FORMAT Values on node2 Executing query SELECT * FROM s3_test order by id FORMAT Values on node2 Executing query SELECT 1 FORMAT Null; SET log_queries = 0; SYSTEM FLUSH LOGS; SELECT count() FROM system.query_log; on ch1 Executing query SELECT 1 FORMAT Null; SET log_queries = 0; SYSTEM FLUSH LOGS; SELECT count() FROM system.query_log; on ch1 Executing query SELECT * FROM s3_test order by id FORMAT Values on node1 Executing query SELECT * FROM s3_test order by id FORMAT Values on node1 Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query OPTIMIZE TABLE s3_test FINAL on node1 Executing query OPTIMIZE TABLE s3_test FINAL on node1 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Starting new HTTP connection (5): 172.16.3.7:9001 Starting new HTTP connection (5): 172.16.3.7:9001 http://172.16.3.7:9001 "GET / HTTP/1.1" 200 0 http://172.16.3.7:9001 "GET / HTTP/1.1" 200 0 Connected to Minio. Connected to Minio. http://172.16.3.7:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.3.7:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.3.7:9001 "PUT /root HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root HTTP/1.1" 200 0 S3 bucket 'root' created S3 bucket 'root' created http://172.16.3.7:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.3.7:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.3.7:9001 "PUT /root2 HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root2 HTTP/1.1" 200 0 S3 bucket 'root2' created S3 bucket 'root2' created ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml up -d --no-recreate] Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT 1 FORMAT Null; SET log_queries = 0; SYSTEM FLUSH LOGS; SELECT count() FROM system.query_log; on ch1 Executing query SELECT 1 FORMAT Null; SET log_queries = 0; SYSTEM FLUSH LOGS; SELECT count() FROM system.query_log; on ch1 Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 Stderr: Container roottests3cluster-gw5-zoo1-1 Running Stderr: Container roottests3cluster-gw5-zoo1-1 Running Stderr: Container roottests3cluster-gw5-zoo2-1 Running Stderr: Container roottests3cluster-gw5-zoo2-1 Running Stderr: Container roottests3cluster-gw5-proxy1-1 Running Stderr: Container roottests3cluster-gw5-proxy1-1 Running Stderr: Container roottests3cluster-gw5-zoo3-1 Running Stderr: Container roottests3cluster-gw5-zoo3-1 Running Stderr: Container roottests3cluster-gw5-proxy2-1 Running Stderr: Container roottests3cluster-gw5-proxy2-1 Running Stderr: Container roottests3cluster-gw5-s0_1_0-1 Creating Stderr: Container roottests3cluster-gw5-s0_1_0-1 Creating Stderr: Container roottests3cluster-gw5-minio1-1 Running Stderr: Container roottests3cluster-gw5-minio1-1 Running Stderr: Container roottests3cluster-gw5-s0_0_1-1 Creating Stderr: Container roottests3cluster-gw5-s0_0_1-1 Creating Stderr: Container roottests3cluster-gw5-resolver-1 Running Stderr: Container roottests3cluster-gw5-resolver-1 Running Stderr: Container roottests3cluster-gw5-s0_0_0-1 Creating Stderr: Container roottests3cluster-gw5-s0_0_0-1 Creating Stderr: Container roottests3cluster-gw5-s0_1_0-1 Created Stderr: Container roottests3cluster-gw5-s0_1_0-1 Created Stderr: Container roottests3cluster-gw5-s0_0_1-1 Created Stderr: Container roottests3cluster-gw5-s0_0_1-1 Created Stderr: Container roottests3cluster-gw5-s0_0_0-1 Created Stderr: Container roottests3cluster-gw5-s0_0_0-1 Created Stderr: Container roottests3cluster-gw5-s0_1_0-1 Starting Stderr: Container roottests3cluster-gw5-s0_1_0-1 Starting Stderr: Container roottests3cluster-gw5-s0_0_1-1 Starting Stderr: Container roottests3cluster-gw5-s0_0_1-1 Starting Stderr: Container roottests3cluster-gw5-s0_0_0-1 Starting Stderr: Container roottests3cluster-gw5-s0_0_0-1 Starting Stderr: Container roottests3cluster-gw5-s0_1_0-1 Started Stderr: Container roottests3cluster-gw5-s0_1_0-1 Started Stderr: Container roottests3cluster-gw5-s0_0_0-1 Started Stderr: Container roottests3cluster-gw5-s0_0_0-1 Started Stderr: Container roottests3cluster-gw5-s0_0_1-1 Started Stderr: Container roottests3cluster-gw5-s0_0_1-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=s0_0_0 get_instance_ip instance_name=s0_0_0 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_0-1/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 get_instance_ip instance_name=s0_0_0 get_instance_ip instance_name=s0_0_0 Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_0-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in s0_0_0, ip: 172.16.3.11... Waiting for ClickHouse start in s0_0_0, ip: 172.16.3.11... http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN QUERIES DEFAULT on main_node Executing query SYSTEM STOP LISTEN QUERIES DEFAULT on main_node http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a106919e940a771f493c98c81c7b164850f731d33e4d5bd316f1b3328e0a2f31/json HTTP/1.1" 200 None ClickHouse s0_0_0 started ClickHouse s0_0_0 started get_instance_ip instance_name=s0_0_1 get_instance_ip instance_name=s0_0_1 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=s0_0_1 get_instance_ip instance_name=s0_0_1 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in s0_0_1, ip: 172.16.3.10... Waiting for ClickHouse start in s0_0_1, ip: 172.16.3.10... http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_0_1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19ae9245deab68c1890fb947da21011a0a18040949129f79bb86553ad8fea008/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19ae9245deab68c1890fb947da21011a0a18040949129f79bb86553ad8fea008/json HTTP/1.1" 200 None ClickHouse s0_0_1 started ClickHouse s0_0_1 started get_instance_ip instance_name=s0_1_0 get_instance_ip instance_name=s0_1_0 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_1_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_1_0-1/json HTTP/1.1" 200 None get_instance_ip instance_name=s0_1_0 get_instance_ip instance_name=s0_1_0 http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_1_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_1_0-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in s0_1_0, ip: 172.16.3.9... Waiting for ClickHouse start in s0_1_0, ip: 172.16.3.9... http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_1_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3cluster-gw5-s0_1_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/47b2f1c6301013c91d316af1208f43cb4cfa02367412041f5e25f5dc32d9bf10/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/47b2f1c6301013c91d316af1208f43cb4cfa02367412041f5e25f5dc32d9bf10/json HTTP/1.1" 200 None ClickHouse s0_1_0 started ClickHouse s0_1_0 started Cluster started Cluster started Executing query SELECT 1 FORMAT Null; SET log_queries = 0; SYSTEM FLUSH LOGS; SELECT count() FROM system.query_log; on ch1 Executing query SELECT 1 FORMAT Null; SET log_queries = 0; SYSTEM FLUSH LOGS; SELECT count() FROM system.query_log; on ch1 http://172.16.3.7:9001 "PUT /root/data/clickhouse/part1.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/clickhouse/part1.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/clickhouse/part123.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/clickhouse/part123.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/database/part2.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/database/part2.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/database/partition675.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/database/partition675.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_0.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_0.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_1.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_1.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_2.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_2.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_3.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_3.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_4.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_4.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_5.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_5.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_6.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_6.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_7.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_7.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_8.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_8.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_9.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_9.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_10.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_10.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_11.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_11.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_12.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_12.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_13.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_13.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_14.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_14.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_15.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_15.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_16.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_16.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_17.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_17.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_18.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_18.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_19.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_19.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_20.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_20.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_21.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_21.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_22.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_22.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_23.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_23.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_24.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_24.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_25.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_25.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_26.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_26.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_27.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_27.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_28.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_28.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_29.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_29.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_30.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_30.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_31.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_31.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_32.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_32.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_33.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_33.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_34.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_34.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_35.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_35.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_36.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_36.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_37.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_37.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_38.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_38.csv HTTP/1.1" 200 0 Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 Executing query SET log_queries = 0; SYSTEM FLUSH LOGS; TRUNCATE TABLE system.query_log; on ch1 http://172.16.3.7:9001 "PUT /root/data/generated/file_39.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_39.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_40.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_40.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_41.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_41.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_42.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_42.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_43.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_43.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_44.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_44.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_45.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_45.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_46.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_46.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_47.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_47.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_48.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_48.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_49.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_49.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_50.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_50.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_51.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_51.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_52.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_52.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_53.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_53.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_54.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_54.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_55.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_55.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_56.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_56.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_57.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_57.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_58.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_58.csv HTTP/1.1" 200 0 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml stop --timeout 20] [gw7] PASSED test_system_queries/test.py::test_system_flush_logs http://172.16.3.7:9001 "PUT /root/data/generated/file_59.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_59.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_60.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_60.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_61.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_61.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_62.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_62.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_63.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_63.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_64.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_64.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_65.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_65.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_66.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_66.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_67.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_67.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_68.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_68.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_69.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_69.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_70.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_70.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_71.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_71.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_72.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_72.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_73.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_73.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_74.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_74.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_75.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_75.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_76.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_76.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_77.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_77.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_78.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_78.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_79.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_79.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_80.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_80.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_81.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_81.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_82.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_82.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_83.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_83.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_84.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_84.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_85.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_85.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_86.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_86.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_87.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_87.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_88.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_88.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_89.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_89.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_90.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_90.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_91.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_91.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_92.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_92.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_93.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_93.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_94.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_94.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_95.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_95.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_96.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_96.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_97.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_97.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_98.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_98.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_99.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "PUT /root/data/generated/file_99.csv HTTP/1.1" 200 0 http://172.16.3.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 http://172.16.3.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Starting mock server s3_mock.py Starting mock server s3_mock.py run container_id:roottests3cluster-gw5-resolver-1 detach:False nothrow:False cmd: ['bash', '-c', 'mkdir -p $(dirname s3_mock.py) && echo aW1wb3J0IHN5cwoKZnJvbSBib3R0bGUgaW1wb3J0IHJlcXVlc3QsIHJlc3BvbnNlLCByb3V0ZSwgcnVuCgoKQHJvdXRlKCIvPF9idWNrZXQ+LzxfcGF0aDpwYXRoPiIpCmRlZiBzZXJ2ZXIoX2J1Y2tldCwgX3BhdGgpOgogICAgcmVzdWx0ID0gKAogICAgICAgIHJlcXVlc3QuaGVhZGVyc1siTXlDdXN0b21IZWFkZXIiXQogICAgICAgIGlmICJNeUN1c3RvbUhlYWRlciIgaW4gcmVxdWVzdC5oZWFkZXJzCiAgICAgICAgZWxzZSAidW5rbm93biIKICAgICkKICAgIHJlc3BvbnNlLmNvbnRlbnRfdHlwZSA9ICJ0ZXh0L3BsYWluIgogICAgcmVzcG9uc2Uuc2V0X2hlYWRlcigiQ29udGVudC1MZW5ndGgiLCBsZW4ocmVzdWx0KSkKICAgIHJldHVybiByZXN1bHQKCgpAcm91dGUoIi8iKQpkZWYgcGluZygpOgogICAgcmVzcG9uc2UuY29udGVudF90eXBlID0gInRleHQvcGxhaW4iCiAgICByZXNwb25zZS5zZXRfaGVhZGVyKCJDb250ZW50LUxlbmd0aCIsIDIpCiAgICByZXR1cm4gIk9LIgoKCnJ1bihob3N0PSIwLjAuMC4wIiwgcG9ydD1pbnQoc3lzLmFyZ3ZbMV0pKQo= | base64 --decode > s3_mock.py'] run container_id:roottests3cluster-gw5-resolver-1 detach:False nothrow:False cmd: ['bash', '-c', 'mkdir -p $(dirname s3_mock.py) && echo aW1wb3J0IHN5cwoKZnJvbSBib3R0bGUgaW1wb3J0IHJlcXVlc3QsIHJlc3BvbnNlLCByb3V0ZSwgcnVuCgoKQHJvdXRlKCIvPF9idWNrZXQ+LzxfcGF0aDpwYXRoPiIpCmRlZiBzZXJ2ZXIoX2J1Y2tldCwgX3BhdGgpOgogICAgcmVzdWx0ID0gKAogICAgICAgIHJlcXVlc3QuaGVhZGVyc1siTXlDdXN0b21IZWFkZXIiXQogICAgICAgIGlmICJNeUN1c3RvbUhlYWRlciIgaW4gcmVxdWVzdC5oZWFkZXJzCiAgICAgICAgZWxzZSAidW5rbm93biIKICAgICkKICAgIHJlc3BvbnNlLmNvbnRlbnRfdHlwZSA9ICJ0ZXh0L3BsYWluIgogICAgcmVzcG9uc2Uuc2V0X2hlYWRlcigiQ29udGVudC1MZW5ndGgiLCBsZW4ocmVzdWx0KSkKICAgIHJldHVybiByZXN1bHQKCgpAcm91dGUoIi8iKQpkZWYgcGluZygpOgogICAgcmVzcG9uc2UuY29udGVudF90eXBlID0gInRleHQvcGxhaW4iCiAgICByZXNwb25zZS5zZXRfaGVhZGVyKCJDb250ZW50LUxlbmd0aCIsIDIpCiAgICByZXR1cm4gIk9LIgoKCnJ1bihob3N0PSIwLjAuMC4wIiwgcG9ydD1pbnQoc3lzLmFyZ3ZbMV0pKQo= | base64 --decode > s3_mock.py'] Command:[docker exec roottests3cluster-gw5-resolver-1 bash -c mkdir -p $(dirname s3_mock.py) && echo aW1wb3J0IHN5cwoKZnJvbSBib3R0bGUgaW1wb3J0IHJlcXVlc3QsIHJlc3BvbnNlLCByb3V0ZSwgcnVuCgoKQHJvdXRlKCIvPF9idWNrZXQ+LzxfcGF0aDpwYXRoPiIpCmRlZiBzZXJ2ZXIoX2J1Y2tldCwgX3BhdGgpOgogICAgcmVzdWx0ID0gKAogICAgICAgIHJlcXVlc3QuaGVhZGVyc1siTXlDdXN0b21IZWFkZXIiXQogICAgICAgIGlmICJNeUN1c3RvbUhlYWRlciIgaW4gcmVxdWVzdC5oZWFkZXJzCiAgICAgICAgZWxzZSAidW5rbm93biIKICAgICkKICAgIHJlc3BvbnNlLmNvbnRlbnRfdHlwZSA9ICJ0ZXh0L3BsYWluIgogICAgcmVzcG9uc2Uuc2V0X2hlYWRlcigiQ29udGVudC1MZW5ndGgiLCBsZW4ocmVzdWx0KSkKICAgIHJldHVybiByZXN1bHQKCgpAcm91dGUoIi8iKQpkZWYgcGluZygpOgogICAgcmVzcG9uc2UuY29udGVudF90eXBlID0gInRleHQvcGxhaW4iCiAgICByZXNwb25zZS5zZXRfaGVhZGVyKCJDb250ZW50LUxlbmd0aCIsIDIpCiAgICByZXR1cm4gIk9LIgoKCnJ1bihob3N0PSIwLjAuMC4wIiwgcG9ydD1pbnQoc3lzLmFyZ3ZbMV0pKQo= | base64 --decode > s3_mock.py] Command:[docker exec roottests3cluster-gw5-resolver-1 bash -c mkdir -p $(dirname s3_mock.py) && echo aW1wb3J0IHN5cwoKZnJvbSBib3R0bGUgaW1wb3J0IHJlcXVlc3QsIHJlc3BvbnNlLCByb3V0ZSwgcnVuCgoKQHJvdXRlKCIvPF9idWNrZXQ+LzxfcGF0aDpwYXRoPiIpCmRlZiBzZXJ2ZXIoX2J1Y2tldCwgX3BhdGgpOgogICAgcmVzdWx0ID0gKAogICAgICAgIHJlcXVlc3QuaGVhZGVyc1siTXlDdXN0b21IZWFkZXIiXQogICAgICAgIGlmICJNeUN1c3RvbUhlYWRlciIgaW4gcmVxdWVzdC5oZWFkZXJzCiAgICAgICAgZWxzZSAidW5rbm93biIKICAgICkKICAgIHJlc3BvbnNlLmNvbnRlbnRfdHlwZSA9ICJ0ZXh0L3BsYWluIgogICAgcmVzcG9uc2Uuc2V0X2hlYWRlcigiQ29udGVudC1MZW5ndGgiLCBsZW4ocmVzdWx0KSkKICAgIHJldHVybiByZXN1bHQKCgpAcm91dGUoIi8iKQpkZWYgcGluZygpOgogICAgcmVzcG9uc2UuY29udGVudF90eXBlID0gInRleHQvcGxhaW4iCiAgICByZXNwb25zZS5zZXRfaGVhZGVyKCJDb250ZW50LUxlbmd0aCIsIDIpCiAgICByZXR1cm4gIk9LIgoKCnJ1bihob3N0PSIwLjAuMC4wIiwgcG9ydD1pbnQoc3lzLmFyZ3ZbMV0pKQo= | base64 --decode > s3_mock.py] run container_id:roottests3cluster-gw5-resolver-1 detach:True nothrow:False cmd: ['bash', '-c', 'python3 s3_mock.py 8080 >/var/log/resolver/s3_mock.log 2>/var/log/resolver/s3_mock.err.log'] run container_id:roottests3cluster-gw5-resolver-1 detach:True nothrow:False cmd: ['bash', '-c', 'python3 s3_mock.py 8080 >/var/log/resolver/s3_mock.log 2>/var/log/resolver/s3_mock.err.log'] Command:[docker exec roottests3cluster-gw5-resolver-1 bash -c python3 s3_mock.py 8080 >/var/log/resolver/s3_mock.log 2>/var/log/resolver/s3_mock.err.log] Command:[docker exec roottests3cluster-gw5-resolver-1 bash -c python3 s3_mock.py 8080 >/var/log/resolver/s3_mock.log 2>/var/log/resolver/s3_mock.err.log] run container_id:roottests3cluster-gw5-resolver-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] run container_id:roottests3cluster-gw5-resolver-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] Command:[docker exec roottests3cluster-gw5-resolver-1 curl -s http://localhost:8080/] Command:[docker exec roottests3cluster-gw5-resolver-1 curl -s http://localhost:8080/] Exitcode:7 Exitcode:7 Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES DEFAULT on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES DEFAULT on backup_node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN QUERIES CUSTOM on main_node Executing query SYSTEM STOP LISTEN QUERIES CUSTOM on main_node Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES CUSTOM on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES CUSTOM on backup_node run container_id:roottests3cluster-gw5-resolver-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] run container_id:roottests3cluster-gw5-resolver-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] Command:[docker exec roottests3cluster-gw5-resolver-1 curl -s http://localhost:8080/] Command:[docker exec roottests3cluster-gw5-resolver-1 curl -s http://localhost:8080/] Stdout:OK Stdout:OK s3_mock.py answered OK on attempt 2 s3_mock.py answered OK on attempt 2 Mock server s3_mock.py started Mock server s3_mock.py started Executing query DROP TABLE IF EXISTS parallel_insert_select ON CLUSTER 'first_shard' SYNC; on s0_0_0 Executing query DROP TABLE IF EXISTS parallel_insert_select ON CLUSTER 'first_shard' SYNC; on s0_0_0 Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query CREATE TABLE parallel_insert_select ON CLUSTER 'first_shard' (a String, b UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/{shard}/insert_select_with_replicated', '{replica}') ORDER BY (a, b); on s0_0_0 Executing query CREATE TABLE parallel_insert_select ON CLUSTER 'first_shard' (a String, b UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/{shard}/insert_select_with_replicated', '{replica}') ORDER BY (a, b); on s0_0_0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query INSERT INTO parallel_insert_select SELECT * FROM s3Cluster( 'first_shard', 'http://minio1:9001/root/data/generated/*.csv', 'minio', 'minio123', 'CSV' ) SETTINGS parallel_distributed_insert_select=1, use_structure_from_insertion_table_in_table_functions=0; on s0_0_0 Executing query INSERT INTO parallel_insert_select SELECT * FROM s3Cluster( 'first_shard', 'http://minio1:9001/root/data/generated/*.csv', 'minio', 'minio123', 'CSV' ) SETTINGS parallel_distributed_insert_select=1, use_structure_from_insertion_table_in_table_functions=0; on s0_0_0 Stderr: Container rootteststorageurl-gw4-nginx-1 Stopping Stderr: Container rootteststorageurl-gw4-nginx-1 Stopping Stderr: Container rootteststorageurl-gw4-node1-1 Stopping Stderr: Container rootteststorageurl-gw4-node1-1 Stopping Stderr: Container rootteststorageurl-gw4-nginx-1 Stopped Stderr: Container rootteststorageurl-gw4-nginx-1 Stopped Stderr: Container rootteststorageurl-gw4-node1-1 Stopped Stderr: Container rootteststorageurl-gw4-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Starting new HTTP connection (1): 172.16.5.5:8123 Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml down --volumes] Starting new HTTP connection (1): 172.16.5.5:8123 Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/.env --project-name rootteststorageurl-gw4 --file /ClickHouse/tests/integration/test_storage_url/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_nginx.yml down --volumes] http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN QUERIES ALL on main_node Executing query SYSTEM STOP LISTEN QUERIES ALL on main_node ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml up -d --no-recreate] Executing query SYSTEM SYNC REPLICA parallel_insert_select on s0_0_0 Executing query SYSTEM SYNC REPLICA parallel_insert_select on s0_0_0 Executing query SELECT count() FROM s3('http://minio1:9001/root/data/generated/*.csv', 'minio', 'minio123', 'CSV','a String, b UInt64') on s0_0_0 Executing query SELECT count() FROM s3('http://minio1:9001/root/data/generated/*.csv', 'minio', 'minio123', 'CSV','a String, b UInt64') on s0_0_0 Executing query SELECT count() FROM parallel_insert_select on s0_0_0 Executing query SELECT count() FROM parallel_insert_select on s0_0_0 [gw5] PASSED test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all Executing query SELECT * from s3( 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') ORDER BY (name, value, polygon) on s0_0_0 Executing query SELECT * from s3( 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') ORDER BY (name, value, polygon) on s0_0_0 Stderr: Container rootteststorageredis-gw1-redis1-1 Running Stderr: Container rootteststorageredis-gw1-redis1-1 Running Stderr: Container rootteststorageredis-gw1-node-1 Creating Stderr: Container rootteststorageredis-gw1-node-1 Creating Stderr: Container rootteststorageredis-gw1-node-1 Created Stderr: Container rootteststorageredis-gw1-node-1 Created Stderr: Container rootteststorageredis-gw1-node-1 Starting Stderr: Container rootteststorageredis-gw1-node-1 Starting Stderr: Container rootteststorageredis-gw1-node-1 Started Stderr: Container rootteststorageredis-gw1-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/rootteststorageredis-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageredis-gw1-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/rootteststorageredis-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageredis-gw1-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.8.3... Waiting for ClickHouse start in node, ip: 172.16.8.3... http://localhost:None "GET /v1.46/containers/rootteststorageredis-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageredis-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None Executing query SELECT * from s3Cluster( 'cluster_simple', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') ORDER BY (name, value, polygon) on s0_0_0 Executing query SELECT * from s3Cluster( 'cluster_simple', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') ORDER BY (name, value, polygon) on s0_0_0 Stderr: Container rootteststorageurl-gw4-node1-1 Stopping Stderr: Container rootteststorageurl-gw4-node1-1 Stopping Stderr: Container rootteststorageurl-gw4-nginx-1 Stopping Stderr: Container rootteststorageurl-gw4-nginx-1 Stopping Stderr: Container rootteststorageurl-gw4-nginx-1 Stopped Stderr: Container rootteststorageurl-gw4-nginx-1 Stopped Stderr: Container rootteststorageurl-gw4-nginx-1 Removing Stderr: Container rootteststorageurl-gw4-nginx-1 Removing Stderr: Container rootteststorageurl-gw4-node1-1 Stopped Stderr: Container rootteststorageurl-gw4-node1-1 Stopped Stderr: Container rootteststorageurl-gw4-node1-1 Removing Stderr: Container rootteststorageurl-gw4-node1-1 Removing Stderr: Container rootteststorageurl-gw4-node1-1 Removed Stderr: Container rootteststorageurl-gw4-node1-1 Removed Stderr: Container rootteststorageurl-gw4-nginx-1 Removed Stderr: Container rootteststorageurl-gw4-nginx-1 Removed Stderr: Network rootteststorageurl-gw4_default Removing Stderr: Network rootteststorageurl-gw4_default Removing Stderr: Network rootteststorageurl-gw4_default Removed Stderr: Network rootteststorageurl-gw4_default Removed Cleanup called Cleanup called Docker networks for project rootteststorageurl-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurl-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststorageurl-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurl-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None Docker volumes for project rootteststorageurl-gw4 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurl-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageurl-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageurl-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageurl-gw4 No running containers for project: rootteststorageurl-gw4 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:12 Stdout:12 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 12 Volumes pruned: 12 test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table] Running tests in /ClickHouse/tests/integration/test_system_detached_tables/test.py Running tests in /ClickHouse/tests/integration/test_system_detached_tables/test.py Cluster start called. is_up=False Cluster start called. is_up=False [gw5] PASSED test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards Executing query SELECT count(*) from s3Cluster( 'cluster_non_existent_port', 'http://minio1:9001/root/data/clickhouse/part1.csv', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') SETTINGS skip_unavailable_shards = 1 on s0_0_0 Executing query SELECT count(*) from s3Cluster( 'cluster_non_existent_port', 'http://minio1:9001/root/data/clickhouse/part1.csv', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') SETTINGS skip_unavailable_shards = 1 on s0_0_0 http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES DEFAULT on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES DEFAULT on backup_node Docker networks for project roottestsystemdetachedtables-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemdetachedtables-gw4 are NETWORK ID NAME DRIVER SCOPE http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Docker containers for project roottestsystemdetachedtables-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemdetachedtables-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemdetachedtables-gw4 are DRIVER VOLUME NAME Docker volumes for project roottestsystemdetachedtables-gw4 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottestsystemdetachedtables-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemdetachedtables-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemdetachedtables-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemdetachedtables-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemdetachedtables-gw4 are DRIVER VOLUME NAME Docker volumes for project roottestsystemdetachedtables-gw4 are DRIVER VOLUME NAME http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None Command:[docker container list --all --filter name='^/roottestsystemdetachedtables-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemdetachedtables-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] [gw5] PASSED test_s3_cluster/test.py::test_skip_unavailable_shards Executing query SELECT * FROM ( SELECT * from s3( 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') UNION ALL SELECT * from s3( 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') ) ORDER BY (name, value, polygon) on s0_0_0 Executing query SELECT * FROM ( SELECT * from s3( 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') UNION ALL SELECT * from s3( 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') ) ORDER BY (name, value, polygon) on s0_0_0 test_s3_cluster/test.py::test_union_all Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemdetachedtables-gw4 No running containers for project: roottestsystemdetachedtables-gw4 Trying to prune unused networks... Trying to prune unused networks... Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:12 Stdout:12 Command:[docker volume prune -f] Command:[docker volume prune -f] http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/668ac8a048624c6da8f55df79978c1ea1d8760eb8291a3f400ff101b8c6c5d7e/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 12 Volumes pruned: 12 Setup directory for instance: replica1 Setup directory for instance: replica1 Executing query DROP TABLE IF EXISTS test_create_table SYNC on node Executing query DROP TABLE IF EXISTS test_create_table SYNC on node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_detached_tables/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_detached_tables/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/database Setup database dir /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/database Setup logs dir /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/logs Setup logs dir /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml pull] Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 Executing query SYSTEM STOP LISTEN QUERIES ALL on main_node Executing query SYSTEM STOP LISTEN QUERIES ALL on main_node Executing query SELECT * FROM ( SELECT * from s3Cluster( 'cluster_simple', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') UNION ALL SELECT * from s3Cluster( 'cluster_simple', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') ) ORDER BY (name, value, polygon) on s0_0_0 Executing query SELECT * FROM ( SELECT * from s3Cluster( 'cluster_simple', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') UNION ALL SELECT * from s3Cluster( 'cluster_simple', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') ) ORDER BY (name, value, polygon) on s0_0_0 Executing query CREATE TABLE test_create_table( k String, v UInt32 ) Engine=Redis('redis1:6379') PRIMARY KEY (k) on node Executing query CREATE TABLE test_create_table( k String, v UInt32 ) Engine=Redis('redis1:6379') PRIMARY KEY (k) on node Executing query DROP TABLE IF EXISTS test_create_table SYNC on node Executing query DROP TABLE IF EXISTS test_create_table SYNC on node [gw5] PASSED test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards Executing query SELECT count(*) from s3Cluster( 'cluster_non_existent_port', 'http://minio1:9001/root/data/clickhouse/part1.csv', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') on s0_0_0 Executing query SELECT count(*) from s3Cluster( 'cluster_non_existent_port', 'http://minio1:9001/root/data/clickhouse/part1.csv', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') on s0_0_0 Executing query CREATE TABLE test_create_table( k String, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse', 10) PRIMARY KEY (k) on node Executing query CREATE TABLE test_create_table( k String, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse', 10) PRIMARY KEY (k) on node [gw5] PASSED test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster Executing query SELECT count(*) from s3Cluster( 'non_existent_cluster', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') UNION ALL SELECT count(*) from s3Cluster( 'non_existent_cluster', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') on s0_0_0 Executing query SELECT count(*) from s3Cluster( 'non_existent_cluster', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') UNION ALL SELECT count(*) from s3Cluster( 'non_existent_cluster', 'http://minio1:9001/root/data/{clickhouse,database}/*', 'minio', 'minio123', 'CSV', 'name String, value UInt32, polygon Array(Array(Tuple(Float64, Float64)))') on s0_0_0 Executing query DROP TABLE IF EXISTS test_create_table SYNC on node Executing query DROP TABLE IF EXISTS test_create_table SYNC on node [gw5] PASSED test_s3_cluster/test.py::test_wrong_cluster Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml stop --timeout 20] Executing query CREATE TABLE test_create_table( k String, f String, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse', 10) PRIMARY KEY (k) on node Executing query CREATE TABLE test_create_table( k String, f String, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse', 10) PRIMARY KEY (k) on node Executing query DROP TABLE IF EXISTS test_create_table SYNC on node Executing query DROP TABLE IF EXISTS test_create_table SYNC on node Executing query CREATE TABLE test_create_table( k String, f String, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse', 10) PRIMARY KEY () on node Executing query CREATE TABLE test_create_table( k String, f String, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse', 10) PRIMARY KEY () on node Executing query DROP TABLE IF EXISTS test_create_table SYNC on node Executing query DROP TABLE IF EXISTS test_create_table SYNC on node Executing query CREATE TABLE test_create_table( k String, f String, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse', 10) on node Executing query CREATE TABLE test_create_table( k String, f String, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse', 10) on node [gw1] PASSED test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete Executing query DROP TABLE IF EXISTS test_delete SYNC on node Executing query DROP TABLE IF EXISTS test_delete SYNC on node Executing query CREATE TABLE test_delete( k UInt32, m DateTime, n String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query CREATE TABLE test_delete( k UInt32, m DateTime, n String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query INSERT INTO test_delete Values (1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy') on node Executing query INSERT INTO test_delete Values (1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy') on node Executing query ALTER TABLE test_delete DELETE WHERE k=1 on node Executing query ALTER TABLE test_delete DELETE WHERE k=1 on node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query SELECT k, m, n FROM test_delete FORMAT TSV on node Executing query SELECT k, m, n FROM test_delete FORMAT TSV on node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES CUSTOM on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES CUSTOM on backup_node Executing query ALTER TABLE test_delete DELETE WHERE m='2023-06-02 00:00:00' on node Executing query ALTER TABLE test_delete DELETE WHERE m='2023-06-02 00:00:00' on node Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT k, m, n FROM test_delete FORMAT TSV on node Executing query SELECT k, m, n FROM test_delete FORMAT TSV on node Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node [gw1] PASSED test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int Executing query DROP TABLE IF EXISTS test_select_int SYNC on node Executing query DROP TABLE IF EXISTS test_select_int SYNC on node Executing query CREATE TABLE test_select_int( k UInt32, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query CREATE TABLE test_select_int( k UInt32, v UInt32 ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query SELECT k, v FROM test_select_int WHERE k=0 FORMAT TSV on node Executing query SELECT k, v FROM test_select_int WHERE k=0 FORMAT TSV on node Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Stderr: Container roottestsystemqueries-gw7-ch1-1 Stopping Stderr: Container roottestsystemqueries-gw7-ch1-1 Stopping Stderr: Container roottestsystemqueries-gw7-ch1-1 Stopped Stderr: Container roottestsystemqueries-gw7-ch1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query SELECT * FROM test_select_int ORDER BY k FORMAT TSV on node Executing query SELECT * FROM test_select_int ORDER BY k FORMAT TSV on node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/.env --project-name roottestsystemqueries-gw7 --file /ClickHouse/tests/integration/test_system_queries/_instances-0-gw7/ch1/docker-compose.yml down --volumes] [gw1] PASSED test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert Executing query DROP TABLE IF EXISTS test_simple_insert SYNC on node Executing query DROP TABLE IF EXISTS test_simple_insert SYNC on node Executing query SELECT * FROM system.merges WHERE table = 'merge_simple' and progress < 1 on node1 Executing query SELECT * FROM system.merges WHERE table = 'merge_simple' and progress < 1 on node1 Executing query CREATE TABLE test_simple_insert( k UInt32, m DateTime, n String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query CREATE TABLE test_simple_insert( k UInt32, m DateTime, n String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query SELECT * FROM system.merges WHERE table = 'merge_simple' and progress < 1 on node1 Executing query SELECT * FROM system.merges WHERE table = 'merge_simple' and progress < 1 on node1 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None [gw9] PASSED test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query INSERT INTO test_simple_insert Values (1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy') on node Executing query INSERT INTO test_simple_insert Values (1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy') on node Executing query DROP TABLE test.merge_simple on node1 Executing query DROP TABLE test.merge_simple on node1 Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT COUNT(*) FROM test_simple_insert FORMAT Values on node Executing query SELECT COUNT(*) FROM test_simple_insert FORMAT Values on node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 [gw6] PASSED test_system_merges/test.py::test_merge_simple[] test_system_merges/test.py::test_merge_simple[replicated] Executing query create table test.merge_simple (a Int64) engine=ReplicatedMergeTree('/clickhouse/test_merge_simple', '{replica}') order by tuple() on node1 Executing query create table test.merge_simple (a Int64) engine=ReplicatedMergeTree('/clickhouse/test_merge_simple', '{replica}') order by tuple() on node1 Executing query SELECT k, m, n FROM test_simple_insert WHERE k=1 FORMAT TSV on node Executing query SELECT k, m, n FROM test_simple_insert WHERE k=1 FORMAT TSV on node Executing query create table test.merge_simple (a Int64) engine=ReplicatedMergeTree('/clickhouse/test_merge_simple', '{replica}') order by tuple() on node2 Executing query create table test.merge_simple (a Int64) engine=ReplicatedMergeTree('/clickhouse/test_merge_simple', '{replica}') order by tuple() on node2 Executing query SELECT k, m, n FROM test_simple_insert WHERE m='2023-06-01 00:00:00' FORMAT TSV on node Executing query SELECT k, m, n FROM test_simple_insert WHERE m='2023-06-01 00:00:00' FORMAT TSV on node Executing query INSERT INTO test.merge_simple VALUES (1) on node1 Executing query INSERT INTO test.merge_simple VALUES (1) on node1 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Executing query SELECT k, m, n FROM test_simple_insert WHERE n='lili' FORMAT TSV on node Executing query SELECT k, m, n FROM test_simple_insert WHERE n='lili' FORMAT TSV on node Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN CUSTOM 'tcp' on main_node Executing query SYSTEM STOP LISTEN CUSTOM 'tcp' on main_node Executing query INSERT INTO test.merge_simple VALUES (2) on node1 Executing query INSERT INTO test.merge_simple VALUES (2) on node1 Stderr: Container roottestsystemqueries-gw7-ch1-1 Stopping Stderr: Container roottestsystemqueries-gw7-ch1-1 Stopping Stderr: Container roottestsystemqueries-gw7-ch1-1 Stopped Stderr: Container roottestsystemqueries-gw7-ch1-1 Stopped Stderr: Container roottestsystemqueries-gw7-ch1-1 Removing Stderr: Container roottestsystemqueries-gw7-ch1-1 Removing Stderr: Container roottestsystemqueries-gw7-ch1-1 Removed Stderr: Container roottestsystemqueries-gw7-ch1-1 Removed Stderr: Network roottestsystemqueries-gw7_default Removing Stderr: Network roottestsystemqueries-gw7_default Removing Stderr: Network roottestsystemqueries-gw7_default Removed Stderr: Network roottestsystemqueries-gw7_default Removed Cleanup called Cleanup called Docker networks for project roottestsystemqueries-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemqueries-gw7 are NETWORK ID NAME DRIVER SCOPE [gw1] PASSED test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select Executing query DROP TABLE IF EXISTS test_simple_select SYNC on node Executing query DROP TABLE IF EXISTS test_simple_select SYNC on node Docker containers for project roottestsystemqueries-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemqueries-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemqueries-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemqueries-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemqueries-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemqueries-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query INSERT INTO test.merge_simple VALUES (3) on node1 Executing query INSERT INTO test.merge_simple VALUES (3) on node1 Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemqueries-gw7 No running containers for project: roottestsystemqueries-gw7 Trying to prune unused networks... Trying to prune unused networks... Executing query CREATE TABLE test_simple_select( k String, v String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query CREATE TABLE test_simple_select( k String, v String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:12 Executing query alter table test.merge_simple add column b int materialized sleepEachRow(3) on node1 Stdout:12 Executing query alter table test.merge_simple add column b int materialized sleepEachRow(3) on node1 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 12 Volumes pruned: 12 test_throttling/test.py::test_backup_throttling[no_local_throttling] Running tests in /ClickHouse/tests/integration/test_throttling/test.py Running tests in /ClickHouse/tests/integration/test_throttling/test.py Cluster start called. is_up=False Cluster start called. is_up=False Executing query SYSTEM START LISTEN CUSTOM 'tcp' on main_node Executing query SYSTEM START LISTEN CUSTOM 'tcp' on main_node Executing query SELECT k, v FROM test_simple_select WHERE k='0' FORMAT TSV on node Executing query SELECT k, v FROM test_simple_select WHERE k='0' FORMAT TSV on node Docker networks for project roottestthrottling-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestthrottling-gw7 are NETWORK ID NAME DRIVER SCOPE Executing query OPTIMIZE TABLE test.merge_simple on node1 Executing query OPTIMIZE TABLE test.merge_simple on node1 Executing query select count() from system.merges where table='merge_simple' on node2 Executing query select count() from system.merges where table='merge_simple' on node2 Docker containers for project roottestthrottling-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestthrottling-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestthrottling-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestthrottling-gw7 are DRIVER VOLUME NAME Cleanup called Cleanup called Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Docker networks for project roottestthrottling-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestthrottling-gw7 are NETWORK ID NAME DRIVER SCOPE Executing query SELECT * FROM test_simple_select ORDER BY k FORMAT TSV on node Executing query SELECT * FROM test_simple_select ORDER BY k FORMAT TSV on node Docker containers for project roottestthrottling-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestthrottling-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestthrottling-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestthrottling-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestthrottling-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestthrottling-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestthrottling-gw7 No running containers for project: roottestthrottling-gw7 Trying to prune unused networks... Trying to prune unused networks... Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node [gw1] PASSED test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Executing query DROP TABLE IF EXISTS test_truncate SYNC on node Executing query DROP TABLE IF EXISTS test_truncate SYNC on node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:12 Stdout:12 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 12 Volumes pruned: 12 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_throttling/configs/static_overrides.xml', '/ClickHouse/tests/integration/test_throttling/configs/dynamic_overrides.xml', '/ClickHouse/tests/integration/test_throttling/configs/ssl.xml'] to /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_throttling/configs/static_overrides.xml', '/ClickHouse/tests/integration/test_throttling/configs/dynamic_overrides.xml', '/ClickHouse/tests/integration/test_throttling/configs/ssl.xml'] to /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/database Setup database dir /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/database Setup logs dir /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/logs Setup logs dir /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_throttling/_instances-0-gw7/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_throttling/_instances-0-gw7/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_throttling/_instances-0-gw7/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_throttling/_instances-0-gw7/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_throttling/_instances-0-gw7/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_throttling/_instances-0-gw7/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_throttling/_instances-0-gw7/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_throttling/_instances-0-gw7/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Executing query CREATE TABLE test_truncate( k UInt32, m DateTime, n String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query CREATE TABLE test_truncate( k UInt32, m DateTime, n String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] Executing query select count() from system.merges where table='merge_simple' on node2 Executing query select count() from system.merges where table='merge_simple' on node2 Executing query INSERT INTO test_truncate Values (1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy') on node Executing query INSERT INTO test_truncate Values (1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy') on node Executing query SELECT database, table, num_parts, source_part_names, source_part_paths, result_part_name, result_part_path, partition_id, is_mutation FROM system.merges WHERE table = 'merge_simple' on node2 Executing query SELECT database, table, num_parts, source_part_names, source_part_paths, result_part_name, result_part_path, partition_id, is_mutation FROM system.merges WHERE table = 'merge_simple' on node2 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN CUSTOM 'http' on main_node Executing query SYSTEM STOP LISTEN CUSTOM 'http' on main_node Executing query TRUNCATE TABLE test_truncate on node Executing query TRUNCATE TABLE test_truncate on node Executing query SELECT COUNT(*) FROM test_truncate FORMAT TSV on node Executing query SELECT COUNT(*) FROM test_truncate FORMAT TSV on node [gw1] PASSED test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update Executing query DROP TABLE IF EXISTS test_update SYNC on node Executing query DROP TABLE IF EXISTS test_update SYNC on node Executing query CREATE TABLE test_update( k UInt32, m DateTime, n String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Executing query CREATE TABLE test_update( k UInt32, m DateTime, n String ) Engine=Redis('redis1:6379', 0, 'clickhouse') PRIMARY KEY (k) on node Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 Executing query SYSTEM START LISTEN CUSTOM 'http' on main_node Executing query SYSTEM START LISTEN CUSTOM 'http' on main_node Executing query INSERT INTO test_update Values (1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy') on node Executing query INSERT INTO test_update Values (1, '2023-06-01 00:00:00', 'lili'), (2, '2023-06-02 00:00:00', 'lucy') on node Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query ALTER TABLE test_update UPDATE m='2023-06-03 00:00:00' WHERE k=1 on node Executing query ALTER TABLE test_update UPDATE m='2023-06-03 00:00:00' WHERE k=1 on node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT k, m, n FROM test_update WHERE k=1 FORMAT TSV on node Executing query SELECT k, m, n FROM test_update WHERE k=1 FORMAT TSV on node Connected to Mongo dbs: ['admin', 'config', 'local'] Connected to Mongo dbs: ['admin', 'config', 'local'] ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate] Executing query ALTER TABLE test_update UPDATE k=2 WHERE k=1 on node Executing query ALTER TABLE test_update UPDATE k=2 WHERE k=1 on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml stop --timeout 20] [gw1] PASSED test_storage_redis/test.py::test_update Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None [gw9] PASSED test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.9.5... Waiting for ClickHouse start in node, ip: 172.16.9.5... http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN TCP on main_node Executing query SYSTEM STOP LISTEN TCP on main_node http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SYSTEM START LISTEN ON CLUSTER default TCP on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default TCP on backup_node http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a57e81309b4214bcd6b501197cf05f0ffeb9f50461c24281900c69c2f48061c1/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Executing query SELECT count() FROM mongodb('mongo_no_cred:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT count() FROM mongodb('mongo_no_cred:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT count() FROM mongodb('mongo_no_cred:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String', options='authSource=admin') on node Executing query SELECT count() FROM mongodb('mongo_no_cred:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String', options='authSource=admin') on node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node [gw3] PASSED test_table_function_mongodb_legacy/test.py::test_auth_source[False] test_table_function_mongodb_legacy/test.py::test_complex_data_type[False] Executing query SELECT COUNT() FROM mongodb('mongo1:27017', 'test', 'complex_table', 'root', 'clickhouse', structure='key UInt64, data String, dict Map(UInt64, String)') on node Executing query SELECT COUNT() FROM mongodb('mongo1:27017', 'test', 'complex_table', 'root', 'clickhouse', structure='key UInt64, data String, dict Map(UInt64, String)') on node Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN HTTP on main_node Executing query SYSTEM STOP LISTEN HTTP on main_node Executing query SELECT sum(key) FROM mongodb('mongo1:27017', 'test', 'complex_table', 'root', 'clickhouse', structure='key UInt64, data String, dict Map(UInt64, String)') on node Executing query SELECT sum(key) FROM mongodb('mongo1:27017', 'test', 'complex_table', 'root', 'clickhouse', structure='key UInt64, data String, dict Map(UInt64, String)') on node Executing query SELECT data from mongodb('mongo1:27017', 'test', 'complex_table', 'root', 'clickhouse', structure='key UInt64, data String, dict Map(UInt64, String)') where key = 42 on node Executing query SELECT data from mongodb('mongo1:27017', 'test', 'complex_table', 'root', 'clickhouse', structure='key UInt64, data String, dict Map(UInt64, String)') where key = 42 on node Stderr: node Pulling Stderr: node Pulling Stderr: node Pulled Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate] Stderr: zoo2 Skipped - Image is already being pulled by replica1 Stderr: zoo2 Skipped - Image is already being pulled by replica1 Stderr: zoo3 Skipped - Image is already being pulled by replica1 Stderr: zoo3 Skipped - Image is already being pulled by replica1 Stderr: zoo1 Skipped - Image is already being pulled by replica1 Stderr: zoo1 Skipped - Image is already being pulled by replica1 Stderr: replica1 Pulling Stderr: replica1 Pulling Stderr: replica1 Pulled Stderr: replica1 Pulled Setup ZooKeeper Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/log', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/config', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/coordination', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/log', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/config', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/coordination', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/log', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/config', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/coordination'] Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/log', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/config', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper1/coordination', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/log', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/config', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper2/coordination', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/log', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/config', '/ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/keeper3/coordination'] Command:[docker compose --project-name roottestsystemdetachedtables-gw4 --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Command:[docker compose --project-name roottestsystemdetachedtables-gw4 --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] [gw3] PASSED test_table_function_mongodb_legacy/test.py::test_complex_data_type[False] test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False] Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 Executing query SYSTEM START LISTEN HTTP on main_node Executing query SYSTEM START LISTEN HTTP on main_node Executing query SELECT aaaa FROM mongodb('mongo1:27017', 'test', 'strange_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT aaaa FROM mongodb('mongo1:27017', 'test', 'strange_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node [gw3] PASSED test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False] test_table_function_mongodb_legacy/test.py::test_missing_columns[False] Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: proxy1 Pulling Stderr: proxy1 Pulling Stderr: minio1 Pulling Stderr: minio1 Pulling Stderr: node Pulling Stderr: node Pulling Stderr: resolver Pulling Stderr: resolver Pulling Stderr: resolver Pulled Stderr: resolver Pulled Stderr: minio1 Pulled Stderr: minio1 Pulled Stderr: proxy1 Pulled Stderr: proxy1 Pulled Stderr: node Pulled Stderr: node Pulled Trying to create Minio instance by command docker compose --project-name roottestthrottling-gw7 --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Trying to create Minio instance by command docker compose --project-name roottestthrottling-gw7 --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Command:[docker compose --project-name roottestthrottling-gw7 --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Command:[docker compose --project-name roottestthrottling-gw7 --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT count() FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data Nullable(String)') WHERE isNull(data) on node Executing query SELECT count() FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data Nullable(String)') WHERE isNull(data) on node [gw3] PASSED test_table_function_mongodb_legacy/test.py::test_missing_columns[False] test_table_function_mongodb_legacy/test.py::test_no_credentials[False] Executing query SELECT count() FROM mongodb('mongo_no_cred:27017', 'test', 'simple_table', '', '', structure='key UInt64, data String') on node Executing query SELECT count() FROM mongodb('mongo_no_cred:27017', 'test', 'simple_table', '', '', structure='key UInt64, data String') on node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 [gw3] PASSED test_table_function_mongodb_legacy/test.py::test_no_credentials[False] test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False] Executing query SELECT count() FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT count() FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN MYSQL on main_node Executing query SYSTEM STOP LISTEN MYSQL on main_node Stderr: Network rootteststorageurlhttpheaders-gw8_default Creating Stderr: Network rootteststorageurlhttpheaders-gw8_default Creating Stderr: Network rootteststorageurlhttpheaders-gw8_default Created Stderr: Network rootteststorageurlhttpheaders-gw8_default Created Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Creating Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Creating Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Created Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Created Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Starting Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Starting Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Started Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/rootteststorageurlhttpheaders-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlhttpheaders-gw8-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/rootteststorageurlhttpheaders-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlhttpheaders-gw8-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.2... Waiting for ClickHouse start in node, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/rootteststorageurlhttpheaders-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlhttpheaders-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml stop --timeout 20] [gw3] PASSED test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False] test_table_function_mongodb_legacy/test.py::test_secure_connection[True] http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None Executing query SYSTEM START LISTEN MYSQL on main_node Executing query SYSTEM START LISTEN MYSQL on main_node http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None Stderr:time="2025-04-02T04:14:37Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:37Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestsystemdetachedtables-gw4_default Creating Stderr: Network roottestsystemdetachedtables-gw4_default Creating Stderr: Network roottestsystemdetachedtables-gw4_default Created Stderr: Network roottestsystemdetachedtables-gw4_default Created Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Creating Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Creating Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Creating Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Creating Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Creating Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Creating Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Created Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Created Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Created Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Created Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Created Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Created Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Starting Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Starting Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Starting Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Starting Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Starting Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Starting Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Started Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Started Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Started Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Started Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Started Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Started Stderr:time="2025-04-02T04:14:38Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:38Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:38Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:38Z" level=debug msg="otel error" error="" Wait ZooKeeper to start Wait ZooKeeper to start get_instance_ip instance_name=zoo1 get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-zoo1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.3, port:2181, use_ssl:False get_kazoo_client: zoo1, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9b33d3fd8bc04434e45b37e254cd1e07213e6a7787198ce56420cb3492c79b22/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'mkdir -p $(dirname /redirect_server.py) && echo aW1wb3J0IGh0dHAuc2VydmVyCmltcG9ydCBzeXMKClJFRElSRUNUX0hPU1QgPSAiIgpSRURJUkVDVF9QT1JUID0gMAoKUkVTVUxUX1BBVEggPSAiL3JlZGlyZWN0X3NlcnZlcl9oZWFkZXJzLnR4dCIKCgpjbGFzcyBSZXF1ZXN0SGFuZGxlcihodHRwLnNlcnZlci5CYXNlSFRUUFJlcXVlc3RIYW5kbGVyKToKICAgIGRlZiBsb2dfbWVzc2FnZShzZWxmLCAqYXJncyk6CiAgICAgICAgd2l0aCBvcGVuKFJFU1VMVF9QQVRILCAidyIpIGFzIGY6CiAgICAgICAgICAgIGYud3JpdGUoc2VsZi5oZWFkZXJzLmFzX3N0cmluZygpKQoKICAgIGRlZiBkb19HRVQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvIjoKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDIwMCkKICAgICAgICAgICAgc2VsZi5zZW5kX2hlYWRlcigiQ29udGVudC1UeXBlIiwgInRleHQvcGxhaW4iKQogICAgICAgICAgICBzZWxmLmVuZF9oZWFkZXJzKCkKICAgICAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZ2xvYmFsIFJFRElSRUNUX0hPU1QsIFJFRElSRUNUX1BPUlQKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDMwMikKICAgICAgICAgICAgdGFyZ2V0X2xvY2F0aW9uID0gZiJodHRwOi8ve1JFRElSRUNUX0hPU1R9OntSRURJUkVDVF9QT1JUfXtzZWxmLnBhdGh9IgogICAgICAgICAgICBzZWxmLnNlbmRfaGVhZGVyKCJMb2NhdGlvbiIsIHRhcmdldF9sb2NhdGlvbikKICAgICAgICAgICAgc2VsZi5lbmRfaGVhZGVycygpCiAgICAgICAgICAgIHNlbGYud2ZpbGUud3JpdGUoYid7InN0YXR1cyI6InJlZGlyZWN0ZWQifScpCgoKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIGhvc3QgPSBzeXMuYXJndlsxXQogICAgcG9ydCA9IGludChzeXMuYXJndlsyXSkKICAgIFJFRElSRUNUX0hPU1QgPSBzeXMuYXJndlszXQogICAgUkVESVJFQ1RfUE9SVCA9IGludChzeXMuYXJndls0XSkKICAgIGh0dHBkID0gaHR0cC5zZXJ2ZXIuVGhyZWFkaW5nSFRUUFNlcnZlcigKICAgICAgICAoCiAgICAgICAgICAgIGhvc3QsCiAgICAgICAgICAgIHBvcnQsCiAgICAgICAgKSwKICAgICAgICBSZXF1ZXN0SGFuZGxlciwKICAgICkKCiAgICB0cnk6CiAgICAgICAgaHR0cGQuc2VydmVfZm9yZXZlcigpCiAgICBmaW5hbGx5OgogICAgICAgIGh0dHBkLnNlcnZlcl9jbG9zZSgpCg== | base64 --decode > /redirect_server.py'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'mkdir -p $(dirname /redirect_server.py) && echo aW1wb3J0IGh0dHAuc2VydmVyCmltcG9ydCBzeXMKClJFRElSRUNUX0hPU1QgPSAiIgpSRURJUkVDVF9QT1JUID0gMAoKUkVTVUxUX1BBVEggPSAiL3JlZGlyZWN0X3NlcnZlcl9oZWFkZXJzLnR4dCIKCgpjbGFzcyBSZXF1ZXN0SGFuZGxlcihodHRwLnNlcnZlci5CYXNlSFRUUFJlcXVlc3RIYW5kbGVyKToKICAgIGRlZiBsb2dfbWVzc2FnZShzZWxmLCAqYXJncyk6CiAgICAgICAgd2l0aCBvcGVuKFJFU1VMVF9QQVRILCAidyIpIGFzIGY6CiAgICAgICAgICAgIGYud3JpdGUoc2VsZi5oZWFkZXJzLmFzX3N0cmluZygpKQoKICAgIGRlZiBkb19HRVQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvIjoKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDIwMCkKICAgICAgICAgICAgc2VsZi5zZW5kX2hlYWRlcigiQ29udGVudC1UeXBlIiwgInRleHQvcGxhaW4iKQogICAgICAgICAgICBzZWxmLmVuZF9oZWFkZXJzKCkKICAgICAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZ2xvYmFsIFJFRElSRUNUX0hPU1QsIFJFRElSRUNUX1BPUlQKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDMwMikKICAgICAgICAgICAgdGFyZ2V0X2xvY2F0aW9uID0gZiJodHRwOi8ve1JFRElSRUNUX0hPU1R9OntSRURJUkVDVF9QT1JUfXtzZWxmLnBhdGh9IgogICAgICAgICAgICBzZWxmLnNlbmRfaGVhZGVyKCJMb2NhdGlvbiIsIHRhcmdldF9sb2NhdGlvbikKICAgICAgICAgICAgc2VsZi5lbmRfaGVhZGVycygpCiAgICAgICAgICAgIHNlbGYud2ZpbGUud3JpdGUoYid7InN0YXR1cyI6InJlZGlyZWN0ZWQifScpCgoKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIGhvc3QgPSBzeXMuYXJndlsxXQogICAgcG9ydCA9IGludChzeXMuYXJndlsyXSkKICAgIFJFRElSRUNUX0hPU1QgPSBzeXMuYXJndlszXQogICAgUkVESVJFQ1RfUE9SVCA9IGludChzeXMuYXJndls0XSkKICAgIGh0dHBkID0gaHR0cC5zZXJ2ZXIuVGhyZWFkaW5nSFRUUFNlcnZlcigKICAgICAgICAoCiAgICAgICAgICAgIGhvc3QsCiAgICAgICAgICAgIHBvcnQsCiAgICAgICAgKSwKICAgICAgICBSZXF1ZXN0SGFuZGxlciwKICAgICkKCiAgICB0cnk6CiAgICAgICAgaHR0cGQuc2VydmVfZm9yZXZlcigpCiAgICBmaW5hbGx5OgogICAgICAgIGh0dHBkLnNlcnZlcl9jbG9zZSgpCg== | base64 --decode > /redirect_server.py'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 bash -c mkdir -p $(dirname /redirect_server.py) && echo aW1wb3J0IGh0dHAuc2VydmVyCmltcG9ydCBzeXMKClJFRElSRUNUX0hPU1QgPSAiIgpSRURJUkVDVF9QT1JUID0gMAoKUkVTVUxUX1BBVEggPSAiL3JlZGlyZWN0X3NlcnZlcl9oZWFkZXJzLnR4dCIKCgpjbGFzcyBSZXF1ZXN0SGFuZGxlcihodHRwLnNlcnZlci5CYXNlSFRUUFJlcXVlc3RIYW5kbGVyKToKICAgIGRlZiBsb2dfbWVzc2FnZShzZWxmLCAqYXJncyk6CiAgICAgICAgd2l0aCBvcGVuKFJFU1VMVF9QQVRILCAidyIpIGFzIGY6CiAgICAgICAgICAgIGYud3JpdGUoc2VsZi5oZWFkZXJzLmFzX3N0cmluZygpKQoKICAgIGRlZiBkb19HRVQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvIjoKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDIwMCkKICAgICAgICAgICAgc2VsZi5zZW5kX2hlYWRlcigiQ29udGVudC1UeXBlIiwgInRleHQvcGxhaW4iKQogICAgICAgICAgICBzZWxmLmVuZF9oZWFkZXJzKCkKICAgICAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZ2xvYmFsIFJFRElSRUNUX0hPU1QsIFJFRElSRUNUX1BPUlQKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDMwMikKICAgICAgICAgICAgdGFyZ2V0X2xvY2F0aW9uID0gZiJodHRwOi8ve1JFRElSRUNUX0hPU1R9OntSRURJUkVDVF9QT1JUfXtzZWxmLnBhdGh9IgogICAgICAgICAgICBzZWxmLnNlbmRfaGVhZGVyKCJMb2NhdGlvbiIsIHRhcmdldF9sb2NhdGlvbikKICAgICAgICAgICAgc2VsZi5lbmRfaGVhZGVycygpCiAgICAgICAgICAgIHNlbGYud2ZpbGUud3JpdGUoYid7InN0YXR1cyI6InJlZGlyZWN0ZWQifScpCgoKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIGhvc3QgPSBzeXMuYXJndlsxXQogICAgcG9ydCA9IGludChzeXMuYXJndlsyXSkKICAgIFJFRElSRUNUX0hPU1QgPSBzeXMuYXJndlszXQogICAgUkVESVJFQ1RfUE9SVCA9IGludChzeXMuYXJndls0XSkKICAgIGh0dHBkID0gaHR0cC5zZXJ2ZXIuVGhyZWFkaW5nSFRUUFNlcnZlcigKICAgICAgICAoCiAgICAgICAgICAgIGhvc3QsCiAgICAgICAgICAgIHBvcnQsCiAgICAgICAgKSwKICAgICAgICBSZXF1ZXN0SGFuZGxlciwKICAgICkKCiAgICB0cnk6CiAgICAgICAgaHR0cGQuc2VydmVfZm9yZXZlcigpCiAgICBmaW5hbGx5OgogICAgICAgIGh0dHBkLnNlcnZlcl9jbG9zZSgpCg== | base64 --decode > /redirect_server.py] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 bash -c mkdir -p $(dirname /redirect_server.py) && echo aW1wb3J0IGh0dHAuc2VydmVyCmltcG9ydCBzeXMKClJFRElSRUNUX0hPU1QgPSAiIgpSRURJUkVDVF9QT1JUID0gMAoKUkVTVUxUX1BBVEggPSAiL3JlZGlyZWN0X3NlcnZlcl9oZWFkZXJzLnR4dCIKCgpjbGFzcyBSZXF1ZXN0SGFuZGxlcihodHRwLnNlcnZlci5CYXNlSFRUUFJlcXVlc3RIYW5kbGVyKToKICAgIGRlZiBsb2dfbWVzc2FnZShzZWxmLCAqYXJncyk6CiAgICAgICAgd2l0aCBvcGVuKFJFU1VMVF9QQVRILCAidyIpIGFzIGY6CiAgICAgICAgICAgIGYud3JpdGUoc2VsZi5oZWFkZXJzLmFzX3N0cmluZygpKQoKICAgIGRlZiBkb19HRVQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvIjoKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDIwMCkKICAgICAgICAgICAgc2VsZi5zZW5kX2hlYWRlcigiQ29udGVudC1UeXBlIiwgInRleHQvcGxhaW4iKQogICAgICAgICAgICBzZWxmLmVuZF9oZWFkZXJzKCkKICAgICAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCiAgICAgICAgZWxzZToKICAgICAgICAgICAgZ2xvYmFsIFJFRElSRUNUX0hPU1QsIFJFRElSRUNUX1BPUlQKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDMwMikKICAgICAgICAgICAgdGFyZ2V0X2xvY2F0aW9uID0gZiJodHRwOi8ve1JFRElSRUNUX0hPU1R9OntSRURJUkVDVF9QT1JUfXtzZWxmLnBhdGh9IgogICAgICAgICAgICBzZWxmLnNlbmRfaGVhZGVyKCJMb2NhdGlvbiIsIHRhcmdldF9sb2NhdGlvbikKICAgICAgICAgICAgc2VsZi5lbmRfaGVhZGVycygpCiAgICAgICAgICAgIHNlbGYud2ZpbGUud3JpdGUoYid7InN0YXR1cyI6InJlZGlyZWN0ZWQifScpCgoKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIGhvc3QgPSBzeXMuYXJndlsxXQogICAgcG9ydCA9IGludChzeXMuYXJndlsyXSkKICAgIFJFRElSRUNUX0hPU1QgPSBzeXMuYXJndlszXQogICAgUkVESVJFQ1RfUE9SVCA9IGludChzeXMuYXJndls0XSkKICAgIGh0dHBkID0gaHR0cC5zZXJ2ZXIuVGhyZWFkaW5nSFRUUFNlcnZlcigKICAgICAgICAoCiAgICAgICAgICAgIGhvc3QsCiAgICAgICAgICAgIHBvcnQsCiAgICAgICAgKSwKICAgICAgICBSZXF1ZXN0SGFuZGxlciwKICAgICkKCiAgICB0cnk6CiAgICAgICAgaHR0cGQuc2VydmVfZm9yZXZlcigpCiAgICBmaW5hbGx5OgogICAgICAgIGh0dHBkLnNlcnZlcl9jbG9zZSgpCg== | base64 --decode > /redirect_server.py] Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:True nothrow:False cmd: ['bash', '-c', 'python3 /redirect_server.py localhost 8080 localhost 8000 > redirect_server.py.log 2>&1'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:True nothrow:False cmd: ['bash', '-c', 'python3 /redirect_server.py localhost 8080 localhost 8000 > redirect_server.py.log 2>&1'] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 bash -c python3 /redirect_server.py localhost 8080 localhost 8000 > redirect_server.py.log 2>&1] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 bash -c python3 /redirect_server.py localhost 8080 localhost 8000 > redirect_server.py.log 2>&1] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8080/] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8080/] Exitcode:7 Exitcode:7 run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8080/] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8080/] Exitcode:7 Exitcode:7 run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8080/'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8080/] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8080/] Stdout:{"status":"ok"} Stdout:{"status":"ok"} run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'mkdir -p $(dirname /http_headers_echo_server.py) && echo aW1wb3J0IGh0dHAuc2VydmVyCmltcG9ydCBqc29uCmltcG9ydCBzeXMKClJFU1VMVF9QQVRIID0gIi9lY2hvX3NlcnZlcl9oZWFkZXJzLnR4dCIKCgpjbGFzcyBSZXF1ZXN0SGFuZGxlcihodHRwLnNlcnZlci5CYXNlSFRUUFJlcXVlc3RIYW5kbGVyKToKICAgIGRlZiBsb2dfbWVzc2FnZShzZWxmLCAqYXJncyk6CiAgICAgICAgd2l0aCBvcGVuKFJFU1VMVF9QQVRILCAidyIpIGFzIGY6CiAgICAgICAgICAgIGYud3JpdGUoc2VsZi5oZWFkZXJzLmFzX3N0cmluZygpKQoKICAgIGRlZiBkb19HRVQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvIjoKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDIwMCkKICAgICAgICAgICAgc2VsZi5zZW5kX2hlYWRlcigiQ29udGVudC1UeXBlIiwgInRleHQvcGxhaW4iKQogICAgICAgICAgICBzZWxmLmVuZF9oZWFkZXJzKCkKICAgICAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvc2FtcGxlLWRhdGEiOgogICAgICAgICAgICBzZWxmLnNlbmRfcmVzcG9uc2UoMjAwKQogICAgICAgICAgICBzZWxmLnNlbmRfaGVhZGVyKCJDb250ZW50LVR5cGUiLCAidGV4dC9wbGFpbiIpCiAgICAgICAgICAgIHNlbGYuZW5kX2hlYWRlcnMoKQogICAgICAgICAgICBzYW1wbGVfZGF0YSA9IFsKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAidGl0bGUiOiAiQ2xpY2tIb3VzZSBOZXdzbGV0dGVyIEp1bmUgMjAyMjogTWF0ZXJpYWxpemVkLCBidXQgc3RpbGwgcmVhbC10aW1lIiwKICAgICAgICAgICAgICAgICAgICAidGhlbWUiOiAiTmV3c2xldHRlciIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJ0aXRsZSI6ICJDbGlja0hvdXNlIE92ZXIgdGhlIFllYXJzIHdpdGggQmVuY2htYXJrcyIsCiAgICAgICAgICAgICAgICAgICAgInRoZW1lIjogIkNsaWNrSG91c2UgSm91cm5leSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICBdCiAgICAgICAgICAgIHNlbGYud2ZpbGUud3JpdGUoYnl0ZXMoanNvbi5kdW1wcyhzYW1wbGVfZGF0YSksICJVVEYtOCIpKQoKICAgIGRlZiBkb19QT1NUKHNlbGYpOgogICAgICAgIHNlbGYucmZpbGUucmVhZDEoKQogICAgICAgIHNlbGYuc2VuZF9yZXNwb25zZSgyMDApCiAgICAgICAgc2VsZi5lbmRfaGVhZGVycygpCiAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCgoKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIGhvc3QgPSBzeXMuYXJndlsxXQogICAgcG9ydCA9IGludChzeXMuYXJndlsyXSkKICAgIGh0dHBkID0gaHR0cC5zZXJ2ZXIuVGhyZWFkaW5nSFRUUFNlcnZlcigKICAgICAgICAoCiAgICAgICAgICAgIGhvc3QsCiAgICAgICAgICAgIHBvcnQsCiAgICAgICAgKSwKICAgICAgICBSZXF1ZXN0SGFuZGxlciwKICAgICkKCiAgICB0cnk6CiAgICAgICAgaHR0cGQuc2VydmVfZm9yZXZlcigpCiAgICBmaW5hbGx5OgogICAgICAgIGh0dHBkLnNlcnZlcl9jbG9zZSgpCg== | base64 --decode > /http_headers_echo_server.py'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'mkdir -p $(dirname /http_headers_echo_server.py) && echo aW1wb3J0IGh0dHAuc2VydmVyCmltcG9ydCBqc29uCmltcG9ydCBzeXMKClJFU1VMVF9QQVRIID0gIi9lY2hvX3NlcnZlcl9oZWFkZXJzLnR4dCIKCgpjbGFzcyBSZXF1ZXN0SGFuZGxlcihodHRwLnNlcnZlci5CYXNlSFRUUFJlcXVlc3RIYW5kbGVyKToKICAgIGRlZiBsb2dfbWVzc2FnZShzZWxmLCAqYXJncyk6CiAgICAgICAgd2l0aCBvcGVuKFJFU1VMVF9QQVRILCAidyIpIGFzIGY6CiAgICAgICAgICAgIGYud3JpdGUoc2VsZi5oZWFkZXJzLmFzX3N0cmluZygpKQoKICAgIGRlZiBkb19HRVQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvIjoKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDIwMCkKICAgICAgICAgICAgc2VsZi5zZW5kX2hlYWRlcigiQ29udGVudC1UeXBlIiwgInRleHQvcGxhaW4iKQogICAgICAgICAgICBzZWxmLmVuZF9oZWFkZXJzKCkKICAgICAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvc2FtcGxlLWRhdGEiOgogICAgICAgICAgICBzZWxmLnNlbmRfcmVzcG9uc2UoMjAwKQogICAgICAgICAgICBzZWxmLnNlbmRfaGVhZGVyKCJDb250ZW50LVR5cGUiLCAidGV4dC9wbGFpbiIpCiAgICAgICAgICAgIHNlbGYuZW5kX2hlYWRlcnMoKQogICAgICAgICAgICBzYW1wbGVfZGF0YSA9IFsKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAidGl0bGUiOiAiQ2xpY2tIb3VzZSBOZXdzbGV0dGVyIEp1bmUgMjAyMjogTWF0ZXJpYWxpemVkLCBidXQgc3RpbGwgcmVhbC10aW1lIiwKICAgICAgICAgICAgICAgICAgICAidGhlbWUiOiAiTmV3c2xldHRlciIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJ0aXRsZSI6ICJDbGlja0hvdXNlIE92ZXIgdGhlIFllYXJzIHdpdGggQmVuY2htYXJrcyIsCiAgICAgICAgICAgICAgICAgICAgInRoZW1lIjogIkNsaWNrSG91c2UgSm91cm5leSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICBdCiAgICAgICAgICAgIHNlbGYud2ZpbGUud3JpdGUoYnl0ZXMoanNvbi5kdW1wcyhzYW1wbGVfZGF0YSksICJVVEYtOCIpKQoKICAgIGRlZiBkb19QT1NUKHNlbGYpOgogICAgICAgIHNlbGYucmZpbGUucmVhZDEoKQogICAgICAgIHNlbGYuc2VuZF9yZXNwb25zZSgyMDApCiAgICAgICAgc2VsZi5lbmRfaGVhZGVycygpCiAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCgoKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIGhvc3QgPSBzeXMuYXJndlsxXQogICAgcG9ydCA9IGludChzeXMuYXJndlsyXSkKICAgIGh0dHBkID0gaHR0cC5zZXJ2ZXIuVGhyZWFkaW5nSFRUUFNlcnZlcigKICAgICAgICAoCiAgICAgICAgICAgIGhvc3QsCiAgICAgICAgICAgIHBvcnQsCiAgICAgICAgKSwKICAgICAgICBSZXF1ZXN0SGFuZGxlciwKICAgICkKCiAgICB0cnk6CiAgICAgICAgaHR0cGQuc2VydmVfZm9yZXZlcigpCiAgICBmaW5hbGx5OgogICAgICAgIGh0dHBkLnNlcnZlcl9jbG9zZSgpCg== | base64 --decode > /http_headers_echo_server.py'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 bash -c mkdir -p $(dirname /http_headers_echo_server.py) && echo aW1wb3J0IGh0dHAuc2VydmVyCmltcG9ydCBqc29uCmltcG9ydCBzeXMKClJFU1VMVF9QQVRIID0gIi9lY2hvX3NlcnZlcl9oZWFkZXJzLnR4dCIKCgpjbGFzcyBSZXF1ZXN0SGFuZGxlcihodHRwLnNlcnZlci5CYXNlSFRUUFJlcXVlc3RIYW5kbGVyKToKICAgIGRlZiBsb2dfbWVzc2FnZShzZWxmLCAqYXJncyk6CiAgICAgICAgd2l0aCBvcGVuKFJFU1VMVF9QQVRILCAidyIpIGFzIGY6CiAgICAgICAgICAgIGYud3JpdGUoc2VsZi5oZWFkZXJzLmFzX3N0cmluZygpKQoKICAgIGRlZiBkb19HRVQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvIjoKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDIwMCkKICAgICAgICAgICAgc2VsZi5zZW5kX2hlYWRlcigiQ29udGVudC1UeXBlIiwgInRleHQvcGxhaW4iKQogICAgICAgICAgICBzZWxmLmVuZF9oZWFkZXJzKCkKICAgICAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvc2FtcGxlLWRhdGEiOgogICAgICAgICAgICBzZWxmLnNlbmRfcmVzcG9uc2UoMjAwKQogICAgICAgICAgICBzZWxmLnNlbmRfaGVhZGVyKCJDb250ZW50LVR5cGUiLCAidGV4dC9wbGFpbiIpCiAgICAgICAgICAgIHNlbGYuZW5kX2hlYWRlcnMoKQogICAgICAgICAgICBzYW1wbGVfZGF0YSA9IFsKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAidGl0bGUiOiAiQ2xpY2tIb3VzZSBOZXdzbGV0dGVyIEp1bmUgMjAyMjogTWF0ZXJpYWxpemVkLCBidXQgc3RpbGwgcmVhbC10aW1lIiwKICAgICAgICAgICAgICAgICAgICAidGhlbWUiOiAiTmV3c2xldHRlciIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJ0aXRsZSI6ICJDbGlja0hvdXNlIE92ZXIgdGhlIFllYXJzIHdpdGggQmVuY2htYXJrcyIsCiAgICAgICAgICAgICAgICAgICAgInRoZW1lIjogIkNsaWNrSG91c2UgSm91cm5leSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICBdCiAgICAgICAgICAgIHNlbGYud2ZpbGUud3JpdGUoYnl0ZXMoanNvbi5kdW1wcyhzYW1wbGVfZGF0YSksICJVVEYtOCIpKQoKICAgIGRlZiBkb19QT1NUKHNlbGYpOgogICAgICAgIHNlbGYucmZpbGUucmVhZDEoKQogICAgICAgIHNlbGYuc2VuZF9yZXNwb25zZSgyMDApCiAgICAgICAgc2VsZi5lbmRfaGVhZGVycygpCiAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCgoKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIGhvc3QgPSBzeXMuYXJndlsxXQogICAgcG9ydCA9IGludChzeXMuYXJndlsyXSkKICAgIGh0dHBkID0gaHR0cC5zZXJ2ZXIuVGhyZWFkaW5nSFRUUFNlcnZlcigKICAgICAgICAoCiAgICAgICAgICAgIGhvc3QsCiAgICAgICAgICAgIHBvcnQsCiAgICAgICAgKSwKICAgICAgICBSZXF1ZXN0SGFuZGxlciwKICAgICkKCiAgICB0cnk6CiAgICAgICAgaHR0cGQuc2VydmVfZm9yZXZlcigpCiAgICBmaW5hbGx5OgogICAgICAgIGh0dHBkLnNlcnZlcl9jbG9zZSgpCg== | base64 --decode > /http_headers_echo_server.py] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 bash -c mkdir -p $(dirname /http_headers_echo_server.py) && echo aW1wb3J0IGh0dHAuc2VydmVyCmltcG9ydCBqc29uCmltcG9ydCBzeXMKClJFU1VMVF9QQVRIID0gIi9lY2hvX3NlcnZlcl9oZWFkZXJzLnR4dCIKCgpjbGFzcyBSZXF1ZXN0SGFuZGxlcihodHRwLnNlcnZlci5CYXNlSFRUUFJlcXVlc3RIYW5kbGVyKToKICAgIGRlZiBsb2dfbWVzc2FnZShzZWxmLCAqYXJncyk6CiAgICAgICAgd2l0aCBvcGVuKFJFU1VMVF9QQVRILCAidyIpIGFzIGY6CiAgICAgICAgICAgIGYud3JpdGUoc2VsZi5oZWFkZXJzLmFzX3N0cmluZygpKQoKICAgIGRlZiBkb19HRVQoc2VsZik6CiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvIjoKICAgICAgICAgICAgc2VsZi5zZW5kX3Jlc3BvbnNlKDIwMCkKICAgICAgICAgICAgc2VsZi5zZW5kX2hlYWRlcigiQ29udGVudC1UeXBlIiwgInRleHQvcGxhaW4iKQogICAgICAgICAgICBzZWxmLmVuZF9oZWFkZXJzKCkKICAgICAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCiAgICAgICAgaWYgc2VsZi5wYXRoID09ICIvc2FtcGxlLWRhdGEiOgogICAgICAgICAgICBzZWxmLnNlbmRfcmVzcG9uc2UoMjAwKQogICAgICAgICAgICBzZWxmLnNlbmRfaGVhZGVyKCJDb250ZW50LVR5cGUiLCAidGV4dC9wbGFpbiIpCiAgICAgICAgICAgIHNlbGYuZW5kX2hlYWRlcnMoKQogICAgICAgICAgICBzYW1wbGVfZGF0YSA9IFsKICAgICAgICAgICAgICAgIHsKICAgICAgICAgICAgICAgICAgICAidGl0bGUiOiAiQ2xpY2tIb3VzZSBOZXdzbGV0dGVyIEp1bmUgMjAyMjogTWF0ZXJpYWxpemVkLCBidXQgc3RpbGwgcmVhbC10aW1lIiwKICAgICAgICAgICAgICAgICAgICAidGhlbWUiOiAiTmV3c2xldHRlciIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICAgICAgewogICAgICAgICAgICAgICAgICAgICJ0aXRsZSI6ICJDbGlja0hvdXNlIE92ZXIgdGhlIFllYXJzIHdpdGggQmVuY2htYXJrcyIsCiAgICAgICAgICAgICAgICAgICAgInRoZW1lIjogIkNsaWNrSG91c2UgSm91cm5leSIsCiAgICAgICAgICAgICAgICB9LAogICAgICAgICAgICBdCiAgICAgICAgICAgIHNlbGYud2ZpbGUud3JpdGUoYnl0ZXMoanNvbi5kdW1wcyhzYW1wbGVfZGF0YSksICJVVEYtOCIpKQoKICAgIGRlZiBkb19QT1NUKHNlbGYpOgogICAgICAgIHNlbGYucmZpbGUucmVhZDEoKQogICAgICAgIHNlbGYuc2VuZF9yZXNwb25zZSgyMDApCiAgICAgICAgc2VsZi5lbmRfaGVhZGVycygpCiAgICAgICAgc2VsZi53ZmlsZS53cml0ZShiJ3sic3RhdHVzIjoib2sifScpCgoKaWYgX19uYW1lX18gPT0gIl9fbWFpbl9fIjoKICAgIGhvc3QgPSBzeXMuYXJndlsxXQogICAgcG9ydCA9IGludChzeXMuYXJndlsyXSkKICAgIGh0dHBkID0gaHR0cC5zZXJ2ZXIuVGhyZWFkaW5nSFRUUFNlcnZlcigKICAgICAgICAoCiAgICAgICAgICAgIGhvc3QsCiAgICAgICAgICAgIHBvcnQsCiAgICAgICAgKSwKICAgICAgICBSZXF1ZXN0SGFuZGxlciwKICAgICkKCiAgICB0cnk6CiAgICAgICAgaHR0cGQuc2VydmVfZm9yZXZlcigpCiAgICBmaW5hbGx5OgogICAgICAgIGh0dHBkLnNlcnZlcl9jbG9zZSgpCg== | base64 --decode > /http_headers_echo_server.py] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:True nothrow:False cmd: ['bash', '-c', 'python3 /http_headers_echo_server.py localhost 8000 > http_headers_echo_server.py.log 2>&1'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:True nothrow:False cmd: ['bash', '-c', 'python3 /http_headers_echo_server.py localhost 8000 > http_headers_echo_server.py.log 2>&1'] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 bash -c python3 /http_headers_echo_server.py localhost 8000 > http_headers_echo_server.py.log 2>&1] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 bash -c python3 /http_headers_echo_server.py localhost 8000 > http_headers_echo_server.py.log 2>&1] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8000/'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8000/'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8000/] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8000/] Exitcode:7 Exitcode:7 run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8000/'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8000/'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8000/] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8000/] Exitcode:7 Exitcode:7 run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8000/'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8000/'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8000/] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8000/] Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Exitcode:7 Exitcode:7 run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8000/'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:True cmd: ['curl', '-s', 'http://localhost:8000/'] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8000/] Command:[docker exec rootteststorageurlhttpheaders-gw8-node-1 curl -s http://localhost:8000/] Stdout:{"status":"ok"} Stdout:{"status":"ok"} Executing query INSERT INTO TABLE FUNCTION url('http://localhost:8000/', JSON, 'a UInt64', headers('X-My-Custom-Header'='test-header')) VALUES (1) on node Executing query INSERT INTO TABLE FUNCTION url('http://localhost:8000/', JSON, 'a UInt64', headers('X-My-Custom-Header'='test-header')) VALUES (1) on node Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN TCP SECURE on main_node Executing query SYSTEM STOP LISTEN TCP SECURE on main_node run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/echo_server_headers.txt'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/echo_server_headers.txt'] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /echo_server_headers.txt] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /echo_server_headers.txt] Stdout:Host: localhost Stdout:Host: localhost Stdout:Transfer-Encoding: chunked Stdout:Transfer-Encoding: chunked Stdout:X-My-Custom-Header: test-header Stdout:X-My-Custom-Header: test-header Stdout:Content-Type: application/json; charset=UTF-8 Stdout:Content-Type: application/json; charset=UTF-8 Stdout:Connection: Keep-Alive Stdout:Connection: Keep-Alive Stdout:Keep-Alive: timeout=10, max=1000 Stdout:Keep-Alive: timeout=10, max=1000 Stdout: Stdout: [gw8] PASSED test_storage_url_http_headers/test.py::test_storage_url_http_headers Executing query SELECT title::String as title, theme::String as theme FROM url('http://127.0.0.1:8080/sample-data', 'JSONEachRow', 'title String, theme String') SETTINGS http_max_tries=2, max_http_get_redirects=2 on node Executing query SELECT title::String as title, theme::String as theme FROM url('http://127.0.0.1:8080/sample-data', 'JSONEachRow', 'title String, theme String') SETTINGS http_max_tries=2, max_http_get_redirects=2 on node test_storage_url_http_headers/test.py::test_storage_url_redirected_headers Stderr: Container rootteststorageredis-gw1-node-1 Stopping Stderr: Container rootteststorageredis-gw1-node-1 Stopping Stderr: Container rootteststorageredis-gw1-redis1-1 Stopping Stderr: Container rootteststorageredis-gw1-redis1-1 Stopping Stderr: Container rootteststorageredis-gw1-redis1-1 Stopped Stderr: Container rootteststorageredis-gw1-redis1-1 Stopped Stderr: Container rootteststorageredis-gw1-node-1 Stopped Stderr: Container rootteststorageredis-gw1-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/.env --project-name rootteststorageredis-gw1 --file /ClickHouse/tests/integration/test_storage_redis/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml down --volumes] Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SYSTEM START LISTEN TCP SECURE on main_node Executing query SYSTEM START LISTEN TCP SECURE on main_node Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/redirect_server_headers.txt'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/redirect_server_headers.txt'] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /redirect_server_headers.txt] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /redirect_server_headers.txt] Stdout:Host: 127.0.0.1 Stdout:Host: 127.0.0.1 Stdout:User-Agent: ClickHouse/25.2.2.96963.altinityantalya.96963 (official build) Stdout:User-Agent: ClickHouse/25.2.2.96963.altinityantalya.96963 (official build) Stdout:Connection: Keep-Alive Stdout:Connection: Keep-Alive Stdout:Keep-Alive: timeout=10, max=1000 Stdout:Keep-Alive: timeout=10, max=1000 Stdout: Stdout: run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/echo_server_headers.txt'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/echo_server_headers.txt'] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /echo_server_headers.txt] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /echo_server_headers.txt] Stdout:Host: localhost Stdout:Host: localhost Stdout:User-Agent: ClickHouse/25.2.2.96963.altinityantalya.96963 (official build) Stdout:User-Agent: ClickHouse/25.2.2.96963.altinityantalya.96963 (official build) Stdout:Connection: Keep-Alive Stdout:Connection: Keep-Alive Stdout:Keep-Alive: timeout=10, max=1000 Stdout:Keep-Alive: timeout=10, max=1000 Stdout: Stdout: [gw8] PASSED test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers Executing query INSERT INTO TABLE FUNCTION url('http://localhost:8000/', JSONEachRow, 'x UInt8') SELECT 1 on node Executing query INSERT INTO TABLE FUNCTION url('http://localhost:8000/', JSONEachRow, 'x UInt8') SELECT 1 on node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Stderr:time="2025-04-02T04:14:38Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:38Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestthrottling-gw7_default Creating Stderr: Network roottestthrottling-gw7_default Creating Stderr: Network roottestthrottling-gw7_default Created Stderr: Network roottestthrottling-gw7_default Created Stderr: Volume "roottestthrottling-gw7_data1-1" Creating Stderr: Volume "roottestthrottling-gw7_data1-1" Creating Stderr: Volume "roottestthrottling-gw7_data1-1" Created Stderr: Volume "roottestthrottling-gw7_data1-1" Created Stderr: Container roottestthrottling-gw7-proxy2-1 Creating Stderr: Container roottestthrottling-gw7-proxy2-1 Creating Stderr: Container roottestthrottling-gw7-proxy1-1 Creating Stderr: Container roottestthrottling-gw7-proxy1-1 Creating Stderr: Container roottestthrottling-gw7-proxy2-1 Created Stderr: Container roottestthrottling-gw7-proxy2-1 Created Stderr: Container roottestthrottling-gw7-proxy1-1 Created Stderr: Container roottestthrottling-gw7-proxy1-1 Created Stderr: Container roottestthrottling-gw7-minio1-1 Creating Stderr: Container roottestthrottling-gw7-minio1-1 Creating Stderr: Container roottestthrottling-gw7-resolver-1 Creating Stderr: Container roottestthrottling-gw7-resolver-1 Creating Stderr: Container roottestthrottling-gw7-resolver-1 Created Stderr: Container roottestthrottling-gw7-resolver-1 Created Stderr: Container roottestthrottling-gw7-minio1-1 Created Stderr: Container roottestthrottling-gw7-minio1-1 Created Stderr: Container roottestthrottling-gw7-proxy2-1 Starting Stderr: Container roottestthrottling-gw7-proxy2-1 Starting Stderr: Container roottestthrottling-gw7-proxy1-1 Starting Stderr: Container roottestthrottling-gw7-proxy1-1 Starting Stderr: Container roottestthrottling-gw7-proxy1-1 Started Stderr: Container roottestthrottling-gw7-proxy1-1 Started Stderr: Container roottestthrottling-gw7-proxy2-1 Started Stderr: Container roottestthrottling-gw7-proxy2-1 Started Stderr: Container roottestthrottling-gw7-minio1-1 Starting Stderr: Container roottestthrottling-gw7-minio1-1 Starting Stderr: Container roottestthrottling-gw7-resolver-1 Starting Stderr: Container roottestthrottling-gw7-resolver-1 Starting Stderr: Container roottestthrottling-gw7-minio1-1 Started Stderr: Container roottestthrottling-gw7-minio1-1 Started Stderr: Container roottestthrottling-gw7-resolver-1 Started Stderr: Container roottestthrottling-gw7-resolver-1 Started Stderr:time="2025-04-02T04:14:40Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:40Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:40Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:40Z" level=debug msg="otel error" error="" Trying to connect to Minio... Trying to connect to Minio... get_instance_ip instance_name=minio1 get_instance_ip instance_name=minio1 run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/echo_server_headers.txt'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/echo_server_headers.txt'] http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-minio1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-minio1-1/json HTTP/1.1" 200 None Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /echo_server_headers.txt] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /echo_server_headers.txt] get_instance_ip instance_name=proxy1 get_instance_ip instance_name=proxy1 http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-proxy1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-proxy1-1/json HTTP/1.1" 200 None Starting new HTTPS connection (1): 172.16.7.4:9001 Starting new HTTPS connection (1): 172.16.7.4:9001 Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTPS connection (2): 172.16.7.4:9001 Starting new HTTPS connection (2): 172.16.7.4:9001 Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTPS connection (3): 172.16.7.4:9001 Starting new HTTPS connection (3): 172.16.7.4:9001 Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTPS connection (4): 172.16.7.4:9001 Starting new HTTPS connection (4): 172.16.7.4:9001 Can't connect to Minio: HTTPSConnectionPool(host='172.16.7.4', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Can't connect to Minio: HTTPSConnectionPool(host='172.16.7.4', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Stdout:Host: localhost Stdout:Host: localhost Stdout:Transfer-Encoding: chunked Stdout:Transfer-Encoding: chunked Stdout:Content-Type: application/x-ndjson; charset=UTF-8 Stdout:Content-Type: application/x-ndjson; charset=UTF-8 Stdout:Connection: Keep-Alive Stdout:Connection: Keep-Alive Stdout:Keep-Alive: timeout=10, max=1000 Stdout:Keep-Alive: timeout=10, max=1000 Stdout: Stdout: Executing query INSERT INTO TABLE FUNCTION url('http://localhost:8000/', JSONEachRow, 'x UInt8', headers('Content-Type' = 'upyachka')) SELECT 1 on node Executing query INSERT INTO TABLE FUNCTION url('http://localhost:8000/', JSONEachRow, 'x UInt8', headers('Content-Type' = 'upyachka')) SELECT 1 on node run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/echo_server_headers.txt'] run container_id:rootteststorageurlhttpheaders-gw8-node-1 detach:False nothrow:False cmd: ['cat', '/echo_server_headers.txt'] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /echo_server_headers.txt] Command:[docker exec -u root rootteststorageurlhttpheaders-gw8-node-1 cat /echo_server_headers.txt] Stdout:Host: localhost Stdout:Host: localhost Stdout:Transfer-Encoding: chunked Stdout:Transfer-Encoding: chunked Stdout:Content-Type: upyachka Stdout:Content-Type: upyachka Stdout:Connection: Keep-Alive Stdout:Connection: Keep-Alive Stdout:Keep-Alive: timeout=10, max=1000 Stdout:Keep-Alive: timeout=10, max=1000 Stdout: Stdout: Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml stop --timeout 20] [gw8] PASSED test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers Stderr: Container rootteststorageredis-gw1-node-1 Stopping Stderr: Container rootteststorageredis-gw1-node-1 Stopping Stderr: Container rootteststorageredis-gw1-redis1-1 Stopping Stderr: Container rootteststorageredis-gw1-redis1-1 Stopping Stderr: Container rootteststorageredis-gw1-node-1 Stopped Stderr: Container rootteststorageredis-gw1-node-1 Stopped Stderr: Container rootteststorageredis-gw1-node-1 Removing Stderr: Container rootteststorageredis-gw1-node-1 Removing Stderr: Container rootteststorageredis-gw1-redis1-1 Stopped Stderr: Container rootteststorageredis-gw1-redis1-1 Stopped Stderr: Container rootteststorageredis-gw1-redis1-1 Removing Stderr: Container rootteststorageredis-gw1-redis1-1 Removing Stderr: Container rootteststorageredis-gw1-node-1 Removed Stderr: Container rootteststorageredis-gw1-node-1 Removed Stderr: Container rootteststorageredis-gw1-redis1-1 Removed Stderr: Container rootteststorageredis-gw1-redis1-1 Removed Stderr: Network rootteststorageredis-gw1_default Removing Stderr: Network rootteststorageredis-gw1_default Removing Stderr: Network rootteststorageredis-gw1_default Removed Stderr: Network rootteststorageredis-gw1_default Removed Cleanup called Cleanup called Docker networks for project rootteststorageredis-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageredis-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststorageredis-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageredis-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None [gw9] PASSED test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Docker volumes for project rootteststorageredis-gw1 are DRIVER VOLUME NAME Docker volumes for project rootteststorageredis-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageredis-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageredis-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageredis-gw1 No running containers for project: rootteststorageredis-gw1 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 test_s3_with_https/test.py::test_s3_with_https[s3_secure] ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 5adb7eee8142 ENV HOSTNAME 5adb7eee8142 ENV SHLVL 0 ENV SHLVL 0 ENV HOME /root ENV HOME /root ENV OLDPWD / ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30050 30051 30052 30053 30054 30055 30056 30057 30058 30059 30060 30061 30062 30063 30064 30065 30066 30067 30068 30069 30070 30071 30072 30073 30074 30075 30076 30077 30078 30079 30080 30081 30082 30083 30084 30085 30086 30087 30088 30089 30090 30091 30092 30093 30094 30095 30096 30097 30098 30099 ENV WORKER_FREE_PORTS 30050 30051 30052 30053 30054 30055 30056 30057 30058 30059 30060 30061 30062 30063 30064 30065 30066 30067 30068 30069 30070 30071 30072 30073 30074 30075 30076 30077 30078 30079 30080 30081 30082 30083 30084 30085 30086 30087 30088 30089 30090 30091 30092 30093 30094 30095 30096 30097 30098 30099 ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_WORKER gw1 ENV PYTEST_XDIST_WORKER gw1 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_s3_with_https/test.py::test_s3_with_https[s3_secure] (setup) ENV PYTEST_CURRENT_TEST test_s3_with_https/test.py::test_s3_with_https[s3_secure] (setup) CLUSTER INIT base_config_dir:/clickhouse-config CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottests3withhttps-gw1. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env', '--project-name', 'roottests3withhttps-gw1', '--file', '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:roottests3withhttps-gw1. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env', '--project-name', 'roottests3withhttps-gw1', '--file', '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Starting cluster... Starting cluster... Running tests in /ClickHouse/tests/integration/test_s3_with_https/test.py Running tests in /ClickHouse/tests/integration/test_s3_with_https/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottests3withhttps-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3withhttps-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottests3withhttps-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3withhttps-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottests3withhttps-gw1 are DRIVER VOLUME NAME Docker volumes for project roottests3withhttps-gw1 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottests3withhttps-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3withhttps-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottests3withhttps-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3withhttps-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottests3withhttps-gw1 are DRIVER VOLUME NAME Docker volumes for project roottests3withhttps-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottests3withhttps-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottests3withhttps-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottests3withhttps-gw1 No running containers for project: roottests3withhttps-gw1 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN QUERIES ALL EXCEPT MYSQL, CUSTOM 'tcp' on main_node Executing query SYSTEM STOP LISTEN QUERIES ALL EXCEPT MYSQL, CUSTOM 'tcp' on main_node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_s3_with_https/configs/config.d/storage_conf.xml', '/ClickHouse/tests/integration/test_s3_with_https/configs/config.d/ssl.xml'] to /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_s3_with_https/configs/config.d/storage_conf.xml', '/ClickHouse/tests/integration/test_s3_with_https/configs/config.d/ssl.xml'] to /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/database Setup database dir /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/database Setup logs dir /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/logs Setup logs dir /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Starting new HTTPS connection (5): 172.16.7.4:9001 Starting new HTTPS connection (5): 172.16.7.4:9001 https://172.16.7.4:9001 "GET / HTTP/1.1" 200 0 https://172.16.7.4:9001 "GET / HTTP/1.1" 200 0 Connected to Minio. Connected to Minio. https://172.16.7.4:9001 "GET /root?location= HTTP/1.1" 404 0 https://172.16.7.4:9001 "GET /root?location= HTTP/1.1" 404 0 https://172.16.7.4:9001 "PUT /root HTTP/1.1" 200 0 https://172.16.7.4:9001 "PUT /root HTTP/1.1" 200 0 S3 bucket 'root' created S3 bucket 'root' created https://172.16.7.4:9001 "GET /root2?location= HTTP/1.1" 404 0 https://172.16.7.4:9001 "GET /root2?location= HTTP/1.1" 404 0 https://172.16.7.4:9001 "PUT /root2 HTTP/1.1" 200 0 https://172.16.7.4:9001 "PUT /root2 HTTP/1.1" 200 0 S3 bucket 'root2' created S3 bucket 'root2' created ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml down --volumes] Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL EXCEPT TCP on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL EXCEPT TCP on backup_node Stderr: Container roottestthrottling-gw7-proxy2-1 Running Stderr: Container roottestthrottling-gw7-proxy2-1 Running Stderr: Container roottestthrottling-gw7-proxy1-1 Running Stderr: Container roottestthrottling-gw7-proxy1-1 Running Stderr: Container roottestthrottling-gw7-minio1-1 Running Stderr: Container roottestthrottling-gw7-minio1-1 Running Stderr: Container roottestthrottling-gw7-resolver-1 Running Stderr: Container roottestthrottling-gw7-resolver-1 Running Stderr: Container roottestthrottling-gw7-node-1 Creating Stderr: Container roottestthrottling-gw7-node-1 Creating Stderr: Container roottestthrottling-gw7-node-1 Created Stderr: Container roottestthrottling-gw7-node-1 Created Stderr: Container roottestthrottling-gw7-node-1 Starting Stderr: Container roottestthrottling-gw7-node-1 Starting Stderr: Container roottestthrottling-gw7-node-1 Started Stderr: Container roottestthrottling-gw7-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.7.6... Waiting for ClickHouse start in node, ip: 172.16.7.6... http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestthrottling-gw7-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cc4924e3ca428bade6ecc4595b3b129ace0b9907eb7665b5938196b08b37182e/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Executing query drop table if exists data; create table data (key UInt64 CODEC(NONE)) engine=MergeTree() order by tuple() settings min_bytes_for_wide_part=1e9, storage_policy='default'; insert into data select * from numbers(1e6); on node Executing query drop table if exists data; create table data (key UInt64 CODEC(NONE)) engine=MergeTree() order by tuple() settings min_bytes_for_wide_part=1e9, storage_policy='default'; insert into data select * from numbers(1e6); on node Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removed Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removing Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removing Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removed Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removed Cleanup called Cleanup called Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttablefunctionmongodblegacy-gw3 No running containers for project: roottesttablefunctionmongodblegacy-gw3 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query backup table data to Disk('default', '1/') on node Executing query backup table data to Disk('default', '1/') on node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:7 Stdout:7 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 7 Volumes pruned: 7 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 5adb7eee8142 ENV HOSTNAME 5adb7eee8142 ENV SHLVL 0 ENV SHLVL 0 ENV HOME /root ENV HOME /root ENV OLDPWD / ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30150 30151 30152 30153 30154 30155 30156 30157 30158 30159 30160 30161 30162 30163 30164 30165 30166 30167 30168 30169 30170 30171 30172 30173 30174 30175 30176 30177 30178 30179 30180 30181 30182 30183 30184 30185 30186 30187 30188 30189 30190 30191 30192 30193 30194 30195 30196 30197 30198 30199 ENV WORKER_FREE_PORTS 30150 30151 30152 30153 30154 30155 30156 30157 30158 30159 30160 30161 30162 30163 30164 30165 30166 30167 30168 30169 30170 30171 30172 30173 30174 30175 30176 30177 30178 30179 30180 30181 30182 30183 30184 30185 30186 30187 30188 30189 30190 30191 30192 30193 30194 30195 30196 30197 30198 30199 ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_WORKER gw3 ENV PYTEST_XDIST_WORKER gw3 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_table_function_mongodb_legacy/test.py::test_secure_connection[True] (setup) ENV PYTEST_CURRENT_TEST test_table_function_mongodb_legacy/test.py::test_secure_connection[True] (setup) CLUSTER INIT base_config_dir:/clickhouse-config CLUSTER INIT base_config_dir:/clickhouse-config Removed :/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3 Removed :/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottesttablefunctionmongodblegacy-gw3. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env', '--project-name', 'roottesttablefunctionmongodblegacy-gw3', '--file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:roottesttablefunctionmongodblegacy-gw3. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env', '--project-name', 'roottesttablefunctionmongodblegacy-gw3', '--file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Running tests in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py Running tests in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Cleanup called Cleanup called run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml] [gw7] PASSED test_throttling/test.py::test_backup_throttling[no_local_throttling] Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/users.d/users_overrides.xml"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/users.d/users_overrides.xml"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/users.d/users_overrides.xml] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/users.d/users_overrides.xml] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttablefunctionmongodblegacy-gw3 No running containers for project: roottesttablefunctionmongodblegacy-gw3 Trying to prune unused networks... Trying to prune unused networks... Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN QUERIES DEFAULT EXCEPT TCP on main_node Executing query SYSTEM STOP LISTEN QUERIES DEFAULT EXCEPT TCP on main_node Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c ps -C clickhouse] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:7 Stdout:7 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:00 clickhouse Stdout: 10 ? 00:00:00 clickhouse run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c pkill clickhouse] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 7 Volumes pruned: 7 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_table_function_mongodb_legacy/mongo_secure_config/config.d/ssl_conf.xml', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/configs/feature_flag.xml'] to /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_table_function_mongodb_legacy/mongo_secure_config/config.d/ssl_conf.xml', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/configs/feature_flag.xml'] to /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/database Setup database dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/database Setup logs dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs Setup logs dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MONGO_HOST': 'mongo1', 'MONGO_EXTERNAL_PORT': '30155', 'MONGO_INTERNAL_PORT': '27017', 'MONGO_NO_CRED_EXTERNAL_PORT': '30156', 'MONGO_NO_CRED_INTERNAL_PORT': '27017', 'MONGO_SECURE_EXTERNAL_PORT': '30157', 'MONGO_SECURE_INTERNAL_PORT': '27017', 'MONGO_SECURE_CONFIG_DIR': '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/mongo_secure_config'} stored in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MONGO_HOST': 'mongo1', 'MONGO_EXTERNAL_PORT': '30155', 'MONGO_INTERNAL_PORT': '27017', 'MONGO_NO_CRED_EXTERNAL_PORT': '30156', 'MONGO_NO_CRED_INTERNAL_PORT': '27017', 'MONGO_SECURE_EXTERNAL_PORT': '30157', 'MONGO_SECURE_INTERNAL_PORT': '27017', 'MONGO_SECURE_CONFIG_DIR': '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/mongo_secure_config'} stored in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml pull] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:10 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES DEFAULT EXCEPT HTTP on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES DEFAULT EXCEPT HTTP on backup_node Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node Stdout:10 Stdout:10 Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Stopping Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Stopping Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Stopped Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/.env --project-name rootteststorageurlhttpheaders-gw8 --file /ClickHouse/tests/integration/test_storage_url_http_headers/_instances-0-gw8/node/docker-compose.yml down --volumes] Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM STOP LISTEN QUERIES CUSTOM EXCEPT CUSTOM 'tcp' on main_node Executing query SYSTEM STOP LISTEN QUERIES CUSTOM EXCEPT CUSTOM 'tcp' on main_node Executing query SELECT * FROM system.merges WHERE table = 'merge_simple' and progress < 1 on node2 Executing query SELECT * FROM system.merges WHERE table = 'merge_simple' and progress < 1 on node2 Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT * FROM system.merges WHERE table = 'merge_simple' and progress < 1 on node2 Executing query SELECT * FROM system.merges WHERE table = 'merge_simple' and progress < 1 on node2 Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Stopping Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Stopping Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Stopped Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Stopped Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Removing Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Removing Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Removed Stderr: Container rootteststorageurlhttpheaders-gw8-node-1 Removed Stderr: Network rootteststorageurlhttpheaders-gw8_default Removing Stderr: Network rootteststorageurlhttpheaders-gw8_default Removing Stderr: Network rootteststorageurlhttpheaders-gw8_default Removed Stderr: Network rootteststorageurlhttpheaders-gw8_default Removed Cleanup called Cleanup called Executing query DROP TABLE test.merge_simple on node1 Executing query DROP TABLE test.merge_simple on node1 Docker networks for project rootteststorageurlhttpheaders-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurlhttpheaders-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststorageurlhttpheaders-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurlhttpheaders-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststorageurlhttpheaders-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurlhttpheaders-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageurlhttpheaders-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageurlhttpheaders-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageurlhttpheaders-gw8 No running containers for project: rootteststorageurlhttpheaders-gw8 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Executing query DROP TABLE test.merge_simple on node2 Executing query DROP TABLE test.merge_simple on node2 Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 Executing query SYSTEM STOP LISTEN QUERIES CUSTOM on main_node Executing query SYSTEM STOP LISTEN QUERIES CUSTOM on main_node Stdout:7 Stdout:7 Command:[docker volume prune -f] Command:[docker volume prune -f] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 7 Volumes pruned: 7 test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side Starting cluster... Starting cluster... Running tests in /ClickHouse/tests/integration/test_server_keep_alive/test.py Running tests in /ClickHouse/tests/integration/test_server_keep_alive/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottestserverkeepalive-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestserverkeepalive-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestserverkeepalive-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestserverkeepalive-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Stdout:10 Stdout:10 Docker volumes for project roottestserverkeepalive-gw8 are DRIVER VOLUME NAME Docker volumes for project roottestserverkeepalive-gw8 are DRIVER VOLUME NAME Cleanup called Cleanup called [gw6] PASSED test_system_merges/test.py::test_merge_simple[replicated] test_system_merges/test.py::test_mutation_simple[] Executing query create table test.mutation_simple (a Int64) engine=MergeTree() order by tuple() on node1 Executing query create table test.mutation_simple (a Int64) engine=MergeTree() order by tuple() on node1 Docker networks for project roottestserverkeepalive-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestserverkeepalive-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestserverkeepalive-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestserverkeepalive-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestserverkeepalive-gw8 are DRIVER VOLUME NAME Docker volumes for project roottestserverkeepalive-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestserverkeepalive-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestserverkeepalive-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestserverkeepalive-gw8 No running containers for project: roottestserverkeepalive-gw8 Trying to prune unused networks... Trying to prune unused networks... Executing query INSERT INTO test.mutation_simple VALUES (1), (2), (3) on node1 Executing query INSERT INTO test.mutation_simple VALUES (1), (2), (3) on node1 Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:7 Stdout:7 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query ALTER TABLE test.mutation_simple UPDATE a = 42 WHERE sleep(9) = 0 on node1 Executing query ALTER TABLE test.mutation_simple UPDATE a = 42 WHERE sleep(9) = 0 on node1 Executing query select count() from system.merges where table='mutation_simple' on node1 Executing query select count() from system.merges where table='mutation_simple' on node1 Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES CUSTOM EXCEPT CUSTOM 'tcp' on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES CUSTOM EXCEPT CUSTOM 'tcp' on backup_node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 7 Volumes pruned: 7 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_server_keep_alive/configs/keep_alive_settings.xml'] to /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_server_keep_alive/configs/keep_alive_settings.xml'] to /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/database Setup database dir /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/database Setup logs dir /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/logs Setup logs dir /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml pull] http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query select count() from system.merges where table='mutation_simple' on node1 Executing query select count() from system.merges where table='mutation_simple' on node1 Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.2.4, port:2181, use_ssl:False get_kazoo_client: zoo2, ip:172.16.2.4, port:2181, use_ssl:False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT database, table, num_parts, source_part_names, source_part_paths, result_part_name, result_part_path, partition_id, is_mutation FROM system.merges WHERE table = 'mutation_simple' on node1 Executing query SELECT database, table, num_parts, source_part_names, source_part_paths, result_part_name, result_part_path, partition_id, is_mutation FROM system.merges WHERE table = 'mutation_simple' on node1 Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-zoo3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.2.2, port:2181, use_ssl:False get_kazoo_client: zoo3, ip:172.16.2.2, port:2181, use_ssl:False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node Executing query SYSTEM START LISTEN ON CLUSTER default QUERIES ALL on backup_node Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate] Executing query SELECT 1 on main_node Executing query SELECT 1 on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Executing query SELECT * FROM mysql('127.0.0.1:9004', 'system', 'one', 'default', '', SETTINGS connect_timeout = 100, connection_wait_timeout = 100) on main_node Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Running Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Running Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Running Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Running Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Running Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Running Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Creating Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Creating Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Created Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Created Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Starting Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Starting Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Started Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=replica1 get_instance_ip instance_name=replica1 http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=replica1 get_instance_ip instance_name=replica1 http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in replica1, ip: 172.16.2.5... Waiting for ClickHouse start in replica1, ip: 172.16.2.5... http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8123 Starting new HTTP connection (1): 172.16.5.5:8123 http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8123 "GET /ping HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:8124 Starting new HTTP connection (1): 172.16.5.5:8124 http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None http://172.16.5.5:8124 "GET /ping HTTP/1.1" 200 None Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml stop --timeout 20] [gw9] PASSED test_system_start_stop_listen/test.py::test_except run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestthrottling-gw7-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestthrottling-gw7-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/79b32e4a6dfbc717e7ee52f7ec17fcdb84fdf7ccd51cd682f5ea2f39e20d5236/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/79b32e4a6dfbc717e7ee52f7ec17fcdb84fdf7ccd51cd682f5ea2f39e20d5236/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/79b32e4a6dfbc717e7ee52f7ec17fcdb84fdf7ccd51cd682f5ea2f39e20d5236/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/79b32e4a6dfbc717e7ee52f7ec17fcdb84fdf7ccd51cd682f5ea2f39e20d5236/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/87b7a5d2fd22ea1836783d68da9934ec31db8ba09e3dd82205bdddc1c0e626a8/json HTTP/1.1" 200 None ClickHouse replica1 started ClickHouse replica1 started Executing query CREATE DATABASE IF NOT EXISTS test_db_atomic ENGINE=Atomic; on replica1 Executing query CREATE DATABASE IF NOT EXISTS test_db_atomic ENGINE=Atomic; on replica1 Executing query CREATE TABLE test_db_atomic.test_table (n Int64) ENGINE=MergeTree ORDER BY n; on replica1 Executing query CREATE TABLE test_db_atomic.test_table (n Int64) ENGINE=MergeTree ORDER BY n; on replica1 Executing query CREATE TABLE test_db_atomic.test_table_perm (n Int64) ENGINE=MergeTree ORDER BY n; on replica1 Executing query CREATE TABLE test_db_atomic.test_table_perm (n Int64) ENGINE=MergeTree ORDER BY n; on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT * FROM system.detached_tables WHERE database='test_db_atomic' on replica1 Executing query SELECT * FROM system.detached_tables WHERE database='test_db_atomic' on replica1 Executing query SET database_replicated_always_detach_permanently=1; DETACH TABLE test_db_atomic.test_table on replica1 Executing query SET database_replicated_always_detach_permanently=1; DETACH TABLE test_db_atomic.test_table on replica1 Executing query DETACH TABLE test_db_atomic.test_table_perm PERMANENTLY on replica1 Executing query DETACH TABLE test_db_atomic.test_table_perm PERMANENTLY on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_atomic' FORMAT Values on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_atomic' FORMAT Values on replica1 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c ps -C clickhouse] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:00 clickhouse Stdout: 10 ? 00:00:00 clickhouse run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c pkill clickhouse] Stdout:780 Stdout:780 Clickhouse process running. Clickhouse process running. run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:780 Stdout:780 Executing query select 20 on node Executing query select 20 on node Stdout:10 Stdout:10 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling] Executing query drop table if exists data; create table data (key UInt64 CODEC(NONE)) engine=MergeTree() order by tuple() settings min_bytes_for_wide_part=1e9, storage_policy='default'; insert into data select * from numbers(1e6); on node Executing query drop table if exists data; create table data (key UInt64 CODEC(NONE)) engine=MergeTree() order by tuple() settings min_bytes_for_wide_part=1e9, storage_policy='default'; insert into data select * from numbers(1e6); on node Executing query backup table data to S3(s3, '2/') on node Executing query backup table data to S3(s3, '2/') on node run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml"] [gw7] PASSED test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/users.d/users_overrides.xml"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/users.d/users_overrides.xml"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/users.d/users_overrides.xml] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/users.d/users_overrides.xml] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 780 ? 00:00:00 clickhouse Stdout: 780 ? 00:00:00 clickhouse run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c pkill clickhouse] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:780 Stdout:780 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Stdout:10 Stdout:10 run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:780 Stdout:780 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:10 run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:780 Stdout:780 Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/.env --project-name roottestsystemstartstoplisten-gw9 --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/main_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_start_stop_listen/_instances-0-gw9/backup_node/docker-compose.yml down --volumes] http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:10 Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-backup_node-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-main_node-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Stopping Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Stopped Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Removing Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-zoo1-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-zoo2-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Removed Stderr: Container roottestsystemstartstoplisten-gw9-zoo3-1 Removed Stderr: Network roottestsystemstartstoplisten-gw9_default Removing Stderr: Network roottestsystemstartstoplisten-gw9_default Removing Stderr: Network roottestsystemstartstoplisten-gw9_default Removed Stderr: Network roottestsystemstartstoplisten-gw9_default Removed Cleanup called Cleanup called Docker networks for project roottestsystemstartstoplisten-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemstartstoplisten-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemstartstoplisten-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemstartstoplisten-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemstartstoplisten-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsystemstartstoplisten-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemstartstoplisten-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemstartstoplisten-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemstartstoplisten-gw9 No running containers for project: roottestsystemstartstoplisten-gw9 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:7 Stdout:7 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 7 Volumes pruned: 7 test_system_logs_recreate/test.py::test_drop_system_log Running tests in /ClickHouse/tests/integration/test_system_logs_recreate/test.py Running tests in /ClickHouse/tests/integration/test_system_logs_recreate/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottestsystemlogsrecreate-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogsrecreate-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemlogsrecreate-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogsrecreate-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemlogsrecreate-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogsrecreate-gw9 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottestsystemlogsrecreate-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogsrecreate-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemlogsrecreate-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogsrecreate-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemlogsrecreate-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogsrecreate-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemlogsrecreate-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemlogsrecreate-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemlogsrecreate-gw9 No running containers for project: roottestsystemlogsrecreate-gw9 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:7 Stdout:7 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 7 Volumes pruned: 7 Setup directory for instance: node_default Setup directory for instance: node_default Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_logs_recreate/configs/config.d/storage_configuration.xml'] to /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_logs_recreate/configs/config.d/storage_configuration.xml'] to /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/configs/config.d run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Setup database dir /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/database Setup database dir /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/database Setup logs dir /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/logs Setup logs dir /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml pull] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestthrottling-gw7-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestthrottling-gw7-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/652a1ea7297ad220aa69e03481c6a3205d12f1459dd3ff9d81af91a50d0cfba8/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/652a1ea7297ad220aa69e03481c6a3205d12f1459dd3ff9d81af91a50d0cfba8/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/652a1ea7297ad220aa69e03481c6a3205d12f1459dd3ff9d81af91a50d0cfba8/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/652a1ea7297ad220aa69e03481c6a3205d12f1459dd3ff9d81af91a50d0cfba8/json HTTP/1.1" 200 586 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/4a58898a60c704f715188982a6e548252f526904f31e1f47e6088f1abd6f0e04/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/4a58898a60c704f715188982a6e548252f526904f31e1f47e6088f1abd6f0e04/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/4a58898a60c704f715188982a6e548252f526904f31e1f47e6088f1abd6f0e04/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/4a58898a60c704f715188982a6e548252f526904f31e1f47e6088f1abd6f0e04/json HTTP/1.1" 200 586 Stderr: mongo1 Skipped - Image is already being pulled by mongo_secure Stderr: mongo1 Skipped - Image is already being pulled by mongo_secure Stderr: mongo_no_cred Skipped - Image is already being pulled by mongo_secure Stderr: mongo_no_cred Skipped - Image is already being pulled by mongo_secure Stderr: node Pulling Stderr: node Pulling Stderr: mongo_secure Pulling Stderr: mongo_secure Pulling Stderr: mongo_secure Pulled Stderr: mongo_secure Pulled Stderr: node Pulled Stderr: node Pulled Setup Mongo Setup Mongo Command:[docker compose --project-name roottesttablefunctionmongodblegacy-gw3 --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml --verbose up -d] Command:[docker compose --project-name roottesttablefunctionmongodblegacy-gw3 --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml --verbose up -d] Stderr: node Pulling Stderr: node Pulling Stderr: node Pulled Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate] Stderr: node_default Pulling Stderr: node_default Pulling Stderr: node_default Pulled Stderr: node_default Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml up -d --no-recreate] Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: node Pulling Stderr: node Pulling Stderr: proxy1 Pulling Stderr: proxy1 Pulling Stderr: minio1 Pulling Stderr: minio1 Pulling Stderr: resolver Pulling Stderr: resolver Pulling Stderr: proxy1 Pulled Stderr: proxy1 Pulled Stderr: node Pulled Stderr: node Pulled Stderr: resolver Pulled Stderr: resolver Pulled Stderr: minio1 Pulled Stderr: minio1 Pulled Trying to create Minio instance by command docker compose --project-name roottests3withhttps-gw1 --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Trying to create Minio instance by command docker compose --project-name roottests3withhttps-gw1 --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Command:[docker compose --project-name roottests3withhttps-gw1 --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Command:[docker compose --project-name roottests3withhttps-gw1 --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1580 Stdout:1580 Clickhouse process running. Clickhouse process running. run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1580 Stdout:1580 Executing query select 20 on node Executing query select 20 on node test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling] Executing query drop table if exists data; create table data (key UInt64 CODEC(NONE)) engine=MergeTree() order by tuple() settings min_bytes_for_wide_part=1e9, storage_policy='s3'; insert into data select * from numbers(1e6); on node Executing query drop table if exists data; create table data (key UInt64 CODEC(NONE)) engine=MergeTree() order by tuple() settings min_bytes_for_wide_part=1e9, storage_policy='s3'; insert into data select * from numbers(1e6); on node http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Stderr: Network roottestserverkeepalive-gw8_default Creating Stderr: Network roottestserverkeepalive-gw8_default Creating Stderr: Network roottestserverkeepalive-gw8_default Created Stderr: Network roottestserverkeepalive-gw8_default Created Stderr: Container roottestserverkeepalive-gw8-node-1 Creating Stderr: Container roottestserverkeepalive-gw8-node-1 Creating Stderr: Container roottestserverkeepalive-gw8-node-1 Created Stderr: Container roottestserverkeepalive-gw8-node-1 Created Stderr: Container roottestserverkeepalive-gw8-node-1 Starting Stderr: Container roottestserverkeepalive-gw8-node-1 Starting Stderr: Container roottestserverkeepalive-gw8-node-1 Started Stderr: Container roottestserverkeepalive-gw8-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestserverkeepalive-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestserverkeepalive-gw8-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestserverkeepalive-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestserverkeepalive-gw8-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.5.2... Waiting for ClickHouse start in node, ip: 172.16.5.2... http://localhost:None "GET /v1.46/containers/roottestserverkeepalive-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestserverkeepalive-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr:time="2025-04-02T04:14:51Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:51Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Creating Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Creating Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Created Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Started Stderr:time="2025-04-02T04:14:52Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:52Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:52Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:52Z" level=debug msg="otel error" error="" Stdout:772 Stdout:772 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None Stdout:772 Stdout:772 Executing query select 20 on replica1 Executing query select 20 on replica1 Executing query backup table data to Disk('default', '3/') on node Executing query backup table data to Disk('default', '3/') on node http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_atomic' FORMAT Values on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_atomic' FORMAT Values on replica1 run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml"] [gw7] PASSED test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/config.d/dynamic_overrides.xml] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/users.d/users_overrides.xml"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '' > /etc/clickhouse-server/users.d/users_overrides.xml"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/users.d/users_overrides.xml] Command:[docker exec roottestthrottling-gw7-node-1 bash -c echo '' > /etc/clickhouse-server/users.d/users_overrides.xml] http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c ps -C clickhouse] Stderr: Network roottestsystemlogsrecreate-gw9_default Creating Stderr: Network roottestsystemlogsrecreate-gw9_default Creating Stderr: Network roottestsystemlogsrecreate-gw9_default Created Stderr: Network roottestsystemlogsrecreate-gw9_default Created Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Creating Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Creating Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Created Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Created Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Starting Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Starting Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Started Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node_default get_instance_ip instance_name=node_default http://localhost:None "GET /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_default get_instance_ip instance_name=node_default http://localhost:None "GET /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_default, ip: 172.16.8.2... Waiting for ClickHouse start in node_default, ip: 172.16.8.2... http://localhost:None "GET /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 1580 ? 00:00:00 clickhouse Stdout: 1580 ? 00:00:00 clickhouse run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestthrottling-gw7-node-1 bash -c pkill clickhouse] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c ps -C clickhouse] http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/de0b9a76ce85976ecf6f88391bbb6f74becd3c642dcf62820ebfe4cd4d8f70d2/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Cluster started Cluster started Starting new HTTP connection (1): 172.16.5.2:8123 Starting new HTTP connection (1): 172.16.5.2:8123 run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_194078 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_194078 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_257991 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_257991 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_272762 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_272762 HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_292935 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_292935 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_466279 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_466279 HTTP/1.1" 200 None Resetting dropped connection: 172.16.5.2 Resetting dropped connection: 172.16.5.2 http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_744869 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_744869 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_88191 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_88191 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_957956 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_957956 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_965747 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_965747 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_993325 HTTP/1.1" 200 None http://172.16.5.2:8123 "GET /?query=select%201&log_comment=test_requests_with_keep_alive_993325 HTTP/1.1" 200 None Executing query SYSTEM FLUSH LOGS; SELECT port FROM system.query_log WHERE log_comment IN ('test_requests_with_keep_alive_194078', 'test_requests_with_keep_alive_257991', 'test_requests_with_keep_alive_272762', 'test_requests_with_keep_alive_292935', 'test_requests_with_keep_alive_466279', 'test_requests_with_keep_alive_744869', 'test_requests_with_keep_alive_88191', 'test_requests_with_keep_alive_957956', 'test_requests_with_keep_alive_965747', 'test_requests_with_keep_alive_993325') AND type = 'QueryFinish' ORDER BY log_comment on node Executing query SYSTEM FLUSH LOGS; SELECT port FROM system.query_log WHERE log_comment IN ('test_requests_with_keep_alive_194078', 'test_requests_with_keep_alive_257991', 'test_requests_with_keep_alive_272762', 'test_requests_with_keep_alive_292935', 'test_requests_with_keep_alive_466279', 'test_requests_with_keep_alive_744869', 'test_requests_with_keep_alive_88191', 'test_requests_with_keep_alive_957956', 'test_requests_with_keep_alive_965747', 'test_requests_with_keep_alive_993325') AND type = 'QueryFinish' ORDER BY log_comment on node Stdout:1580 Stdout:1580 Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 772 ? 00:00:00 clickhouse Stdout: 772 ? 00:00:00 clickhouse run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c pkill clickhouse] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None Stdout:772 Stdout:772 http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None Command:[docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml stop --timeout 20] [gw8] PASSED test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side Stderr:time="2025-04-02T04:14:51Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:14:51Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottests3withhttps-gw1_default Creating Stderr: Network roottests3withhttps-gw1_default Creating Stderr: Network roottests3withhttps-gw1_default Created Stderr: Network roottests3withhttps-gw1_default Created Stderr: Volume "roottests3withhttps-gw1_data1-1" Creating Stderr: Volume "roottests3withhttps-gw1_data1-1" Creating Stderr: Volume "roottests3withhttps-gw1_data1-1" Created Stderr: Volume "roottests3withhttps-gw1_data1-1" Created Stderr: Container roottests3withhttps-gw1-proxy2-1 Creating Stderr: Container roottests3withhttps-gw1-proxy2-1 Creating Stderr: Container roottests3withhttps-gw1-proxy1-1 Creating Stderr: Container roottests3withhttps-gw1-proxy1-1 Creating Stderr: Container roottests3withhttps-gw1-proxy1-1 Created Stderr: Container roottests3withhttps-gw1-proxy1-1 Created Stderr: Container roottests3withhttps-gw1-proxy2-1 Created Stderr: Container roottests3withhttps-gw1-proxy2-1 Created Stderr: Container roottests3withhttps-gw1-minio1-1 Creating Stderr: Container roottests3withhttps-gw1-minio1-1 Creating Stderr: Container roottests3withhttps-gw1-resolver-1 Creating Stderr: Container roottests3withhttps-gw1-resolver-1 Creating Stderr: Container roottests3withhttps-gw1-minio1-1 Created Stderr: Container roottests3withhttps-gw1-minio1-1 Created Stderr: Container roottests3withhttps-gw1-resolver-1 Created Stderr: Container roottests3withhttps-gw1-resolver-1 Created Stderr: Container roottests3withhttps-gw1-proxy2-1 Starting Stderr: Container roottests3withhttps-gw1-proxy2-1 Starting Stderr: Container roottests3withhttps-gw1-proxy1-1 Starting Stderr: Container roottests3withhttps-gw1-proxy1-1 Starting Stderr: Container roottests3withhttps-gw1-proxy1-1 Started Stderr: Container roottests3withhttps-gw1-proxy1-1 Started Stderr: Container roottests3withhttps-gw1-proxy2-1 Started Stderr: Container roottests3withhttps-gw1-proxy2-1 Started Stderr: Container roottests3withhttps-gw1-minio1-1 Starting Stderr: Container roottests3withhttps-gw1-minio1-1 Starting Stderr: Container roottests3withhttps-gw1-resolver-1 Starting Stderr: Container roottests3withhttps-gw1-resolver-1 Starting Stderr: Container roottests3withhttps-gw1-resolver-1 Started Stderr: Container roottests3withhttps-gw1-resolver-1 Started Stderr: Container roottests3withhttps-gw1-minio1-1 Started Stderr: Container roottests3withhttps-gw1-minio1-1 Started Stderr:time="2025-04-02T04:14:53Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:53Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:53Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:14:53Z" level=debug msg="otel error" error="" Trying to connect to Minio... Trying to connect to Minio... get_instance_ip instance_name=minio1 get_instance_ip instance_name=minio1 http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-minio1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-minio1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=proxy1 get_instance_ip instance_name=proxy1 http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-proxy1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-proxy1-1/json HTTP/1.1" 200 None Starting new HTTPS connection (1): 172.16.9.4:9001 Starting new HTTPS connection (1): 172.16.9.4:9001 Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTPS connection (2): 172.16.9.4:9001 Starting new HTTPS connection (2): 172.16.9.4:9001 Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTPS connection (3): 172.16.9.4:9001 Starting new HTTPS connection (3): 172.16.9.4:9001 Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTPS connection (4): 172.16.9.4:9001 Starting new HTTPS connection (4): 172.16.9.4:9001 Can't connect to Minio: HTTPSConnectionPool(host='172.16.9.4', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Can't connect to Minio: HTTPSConnectionPool(host='172.16.9.4', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/19f4d009edca837ecf66864b2acb6af4f7599f576fc55f8d57d8fb89bf6582ad/json HTTP/1.1" 200 None ClickHouse node_default started ClickHouse node_default started run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n 1000000\n \n \n " > /etc/clickhouse-server/config.d/yyy-override-query_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n 1000000\n \n \n " > /etc/clickhouse-server/config.d/yyy-override-query_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " 1000000 " > /etc/clickhouse-server/config.d/yyy-override-query_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " 1000000 " > /etc/clickhouse-server/config.d/yyy-override-query_log.xml ] http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:00 clickhouse Stdout: 10 ? 00:00:00 clickhouse run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:10 run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1580 Stdout:1580 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:772 Stdout:772 Starting new HTTPS connection (5): 172.16.9.4:9001 Starting new HTTPS connection (5): 172.16.9.4:9001 https://172.16.9.4:9001 "GET / HTTP/1.1" 200 0 https://172.16.9.4:9001 "GET / HTTP/1.1" 200 0 Connected to Minio. Connected to Minio. https://172.16.9.4:9001 "GET /root?location= HTTP/1.1" 404 0 https://172.16.9.4:9001 "GET /root?location= HTTP/1.1" 404 0 https://172.16.9.4:9001 "PUT /root HTTP/1.1" 200 0 https://172.16.9.4:9001 "PUT /root HTTP/1.1" 200 0 S3 bucket 'root' created S3 bucket 'root' created https://172.16.9.4:9001 "GET /root2?location= HTTP/1.1" 404 0 https://172.16.9.4:9001 "GET /root2?location= HTTP/1.1" 404 0 https://172.16.9.4:9001 "PUT /root2 HTTP/1.1" 200 0 https://172.16.9.4:9001 "PUT /root2 HTTP/1.1" 200 0 S3 bucket 'root2' created S3 bucket 'root2' created ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Stderr: Container roottests3cluster-gw5-s0_0_1-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_1-1 Stopping Stderr: Container roottests3cluster-gw5-resolver-1 Stopping Stderr: Container roottests3cluster-gw5-resolver-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_0-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_0-1 Stopping Stderr: Container roottests3cluster-gw5-s0_1_0-1 Stopping Stderr: Container roottests3cluster-gw5-s0_1_0-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_1-1 Stopped Stderr: Container roottests3cluster-gw5-s0_0_1-1 Stopped Stderr: Container roottests3cluster-gw5-s0_0_0-1 Stopped Stderr: Container roottests3cluster-gw5-s0_0_0-1 Stopped Stderr: Container roottests3cluster-gw5-minio1-1 Stopping Stderr: Container roottests3cluster-gw5-minio1-1 Stopping Stderr: Container roottests3cluster-gw5-minio1-1 Stopped Stderr: Container roottests3cluster-gw5-minio1-1 Stopped Stderr: Container roottests3cluster-gw5-s0_1_0-1 Stopped Stderr: Container roottests3cluster-gw5-s0_1_0-1 Stopped Stderr: Container roottests3cluster-gw5-zoo1-1 Stopping Stderr: Container roottests3cluster-gw5-zoo1-1 Stopping Stderr: Container roottests3cluster-gw5-zoo3-1 Stopping Stderr: Container roottests3cluster-gw5-zoo3-1 Stopping Stderr: Container roottests3cluster-gw5-zoo2-1 Stopping Stderr: Container roottests3cluster-gw5-zoo2-1 Stopping Stderr: Container roottests3cluster-gw5-zoo2-1 Stopped Stderr: Container roottests3cluster-gw5-zoo2-1 Stopped Stderr: Container roottests3cluster-gw5-zoo3-1 Stopped Stderr: Container roottests3cluster-gw5-zoo3-1 Stopped Stderr: Container roottests3cluster-gw5-zoo1-1 Stopped Stderr: Container roottests3cluster-gw5-zoo1-1 Stopped Stderr: Container roottests3cluster-gw5-resolver-1 Stopped Stderr: Container roottests3cluster-gw5-resolver-1 Stopped Stderr: Container roottests3cluster-gw5-proxy1-1 Stopping Stderr: Container roottests3cluster-gw5-proxy1-1 Stopping Stderr: Container roottests3cluster-gw5-proxy2-1 Stopping Stderr: Container roottests3cluster-gw5-proxy2-1 Stopping Stderr: Container roottests3cluster-gw5-proxy2-1 Stopped Stderr: Container roottests3cluster-gw5-proxy2-1 Stopped Stderr: Container roottests3cluster-gw5-proxy1-1 Stopped Stderr: Container roottests3cluster-gw5-proxy1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/.env --project-name roottests3cluster-gw5 --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_0_1/docker-compose.yml --file /ClickHouse/tests/integration/test_s3_cluster/_instances-0-gw5/s0_1_0/docker-compose.yml down --volumes] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:10 Stderr: Container roottests3withhttps-gw1-proxy1-1 Running Stderr: Container roottests3withhttps-gw1-proxy1-1 Running Stderr: Container roottests3withhttps-gw1-proxy2-1 Running Stderr: Container roottests3withhttps-gw1-proxy2-1 Running Stderr: Container roottests3withhttps-gw1-minio1-1 Running Stderr: Container roottests3withhttps-gw1-minio1-1 Running Stderr: Container roottests3withhttps-gw1-resolver-1 Running Stderr: Container roottests3withhttps-gw1-resolver-1 Running Stderr: Container roottests3withhttps-gw1-node-1 Creating Stderr: Container roottests3withhttps-gw1-node-1 Creating Stderr: Container roottests3withhttps-gw1-node-1 Created Stderr: Container roottests3withhttps-gw1-node-1 Created Stderr: Container roottests3withhttps-gw1-node-1 Starting Stderr: Container roottests3withhttps-gw1-node-1 Starting Stderr: Container roottests3withhttps-gw1-node-1 Started Stderr: Container roottests3withhttps-gw1-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.9.6... Waiting for ClickHouse start in node, ip: 172.16.9.6... http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None Stdout:1580 Stdout:1580 Stderr: Container roottests3cluster-gw5-resolver-1 Stopping Stderr: Container roottests3cluster-gw5-resolver-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_1-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_1-1 Stopping Stderr: Container roottests3cluster-gw5-s0_1_0-1 Stopping Stderr: Container roottests3cluster-gw5-s0_1_0-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_0-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_0-1 Stopping Stderr: Container roottests3cluster-gw5-s0_0_1-1 Stopped Stderr: Container roottests3cluster-gw5-s0_0_1-1 Stopped Stderr: Container roottests3cluster-gw5-s0_0_1-1 Removing Stderr: Container roottests3cluster-gw5-s0_0_1-1 Removing Stderr: Container roottests3cluster-gw5-resolver-1 Stopped Stderr: Container roottests3cluster-gw5-resolver-1 Stopped Stderr: Container roottests3cluster-gw5-resolver-1 Removing Stderr: Container roottests3cluster-gw5-resolver-1 Removing Stderr: Container roottests3cluster-gw5-s0_0_0-1 Stopped Stderr: Container roottests3cluster-gw5-s0_0_0-1 Stopped Stderr: Container roottests3cluster-gw5-s0_0_0-1 Removing Stderr: Container roottests3cluster-gw5-s0_0_0-1 Removing Stderr: Container roottests3cluster-gw5-s0_1_0-1 Stopped Stderr: Container roottests3cluster-gw5-s0_1_0-1 Stopped Stderr: Container roottests3cluster-gw5-s0_1_0-1 Removing Stderr: Container roottests3cluster-gw5-s0_1_0-1 Removing Stderr: Container roottests3cluster-gw5-s0_0_0-1 Removed Stderr: Container roottests3cluster-gw5-s0_0_0-1 Removed Stderr: Container roottests3cluster-gw5-minio1-1 Stopping Stderr: Container roottests3cluster-gw5-minio1-1 Stopping Stderr: Container roottests3cluster-gw5-s0_1_0-1 Removed Stderr: Container roottests3cluster-gw5-s0_1_0-1 Removed Stderr: Container roottests3cluster-gw5-s0_0_1-1 Removed Stderr: Container roottests3cluster-gw5-s0_0_1-1 Removed Stderr: Container roottests3cluster-gw5-zoo3-1 Stopping Stderr: Container roottests3cluster-gw5-zoo3-1 Stopping Stderr: Container roottests3cluster-gw5-zoo1-1 Stopping Stderr: Container roottests3cluster-gw5-zoo1-1 Stopping Stderr: Container roottests3cluster-gw5-zoo2-1 Stopping Stderr: Container roottests3cluster-gw5-zoo2-1 Stopping Stderr: Container roottests3cluster-gw5-minio1-1 Stopped Stderr: Container roottests3cluster-gw5-minio1-1 Stopped Stderr: Container roottests3cluster-gw5-minio1-1 Removing Stderr: Container roottests3cluster-gw5-minio1-1 Removing Stderr: Container roottests3cluster-gw5-zoo3-1 Stopped Stderr: Container roottests3cluster-gw5-zoo3-1 Stopped Stderr: Container roottests3cluster-gw5-zoo3-1 Removing Stderr: Container roottests3cluster-gw5-zoo3-1 Removing Stderr: Container roottests3cluster-gw5-resolver-1 Removed Stderr: Container roottests3cluster-gw5-resolver-1 Removed Stderr: Container roottests3cluster-gw5-zoo2-1 Stopped Stderr: Container roottests3cluster-gw5-zoo2-1 Stopped Stderr: Container roottests3cluster-gw5-zoo2-1 Removing Stderr: Container roottests3cluster-gw5-zoo2-1 Removing Stderr: Container roottests3cluster-gw5-zoo1-1 Stopped Stderr: Container roottests3cluster-gw5-zoo1-1 Stopped Stderr: Container roottests3cluster-gw5-zoo1-1 Removing Stderr: Container roottests3cluster-gw5-zoo1-1 Removing Stderr: Container roottests3cluster-gw5-zoo2-1 Removed Stderr: Container roottests3cluster-gw5-zoo2-1 Removed Stderr: Container roottests3cluster-gw5-zoo1-1 Removed Stderr: Container roottests3cluster-gw5-zoo1-1 Removed Stderr: Container roottests3cluster-gw5-zoo3-1 Removed Stderr: Container roottests3cluster-gw5-zoo3-1 Removed Stderr: Container roottests3cluster-gw5-minio1-1 Removed Stderr: Container roottests3cluster-gw5-minio1-1 Removed Stderr: Container roottests3cluster-gw5-proxy2-1 Stopping Stderr: Container roottests3cluster-gw5-proxy2-1 Stopping Stderr: Container roottests3cluster-gw5-proxy1-1 Stopping Stderr: Container roottests3cluster-gw5-proxy1-1 Stopping Stderr: Container roottests3cluster-gw5-proxy2-1 Stopped Stderr: Container roottests3cluster-gw5-proxy2-1 Stopped Stderr: Container roottests3cluster-gw5-proxy2-1 Removing Stderr: Container roottests3cluster-gw5-proxy2-1 Removing Stderr: Container roottests3cluster-gw5-proxy1-1 Stopped Stderr: Container roottests3cluster-gw5-proxy1-1 Stopped Stderr: Container roottests3cluster-gw5-proxy1-1 Removing Stderr: Container roottests3cluster-gw5-proxy1-1 Removing Stderr: Container roottests3cluster-gw5-proxy1-1 Removed Stderr: Container roottests3cluster-gw5-proxy1-1 Removed Stderr: Container roottests3cluster-gw5-proxy2-1 Removed Stderr: Container roottests3cluster-gw5-proxy2-1 Removed Stderr: Volume roottests3cluster-gw5_data1-1 Removing Stderr: Volume roottests3cluster-gw5_data1-1 Removing Stderr: Network roottests3cluster-gw5_default Removing Stderr: Network roottests3cluster-gw5_default Removing Stderr: Volume roottests3cluster-gw5_data1-1 Removed Stderr: Volume roottests3cluster-gw5_data1-1 Removed Stderr: Network roottests3cluster-gw5_default Removed Stderr: Network roottests3cluster-gw5_default Removed run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Cleanup called Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Cleanup called Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Docker networks for project roottests3cluster-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3cluster-gw5 are NETWORK ID NAME DRIVER SCOPE http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None Docker containers for project roottests3cluster-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3cluster-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Stdout:772 Stdout:772 Docker volumes for project roottests3cluster-gw5 are DRIVER VOLUME NAME Docker volumes for project roottests3cluster-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottests3cluster-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottests3cluster-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottests3cluster-gw5 No running containers for project: roottests3cluster-gw5 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0350cbc8d2784fee79eb393dcff3943c0225b34b54ab583ca527e07098c69f3d/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Cluster started Cluster started Executing query CREATE TABLE s3_test ( id Int64, data String ) ENGINE=MergeTree() ORDER BY id SETTINGS storage_policy='s3_secure' on node Executing query CREATE TABLE s3_test ( id Int64, data String ) ENGINE=MergeTree() ORDER BY id SETTINGS storage_policy='s3_secure' on node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 test_s3_table_functions/test.py::test_s3_table_functions Starting cluster... Starting cluster... Running tests in /ClickHouse/tests/integration/test_s3_table_functions/test.py Running tests in /ClickHouse/tests/integration/test_s3_table_functions/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottests3tablefunctions-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3tablefunctions-gw5 are NETWORK ID NAME DRIVER SCOPE http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Docker containers for project roottests3tablefunctions-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3tablefunctions-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query INSERT INTO s3_test VALUES (0,'data'),(1,'data') on node Executing query INSERT INTO s3_test VALUES (0,'data'),(1,'data') on node Docker volumes for project roottests3tablefunctions-gw5 are DRIVER VOLUME NAME Docker volumes for project roottests3tablefunctions-gw5 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottests3tablefunctions-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3tablefunctions-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottests3tablefunctions-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3tablefunctions-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottests3tablefunctions-gw5 are DRIVER VOLUME NAME Docker volumes for project roottests3tablefunctions-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottests3tablefunctions-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottests3tablefunctions-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottests3tablefunctions-gw5 No running containers for project: roottests3tablefunctions-gw5 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stderr: Container roottestserverkeepalive-gw8-node-1 Stopping Stderr: Container roottestserverkeepalive-gw8-node-1 Stopping Stderr: Container roottestserverkeepalive-gw8-node-1 Stopped Stderr: Container roottestserverkeepalive-gw8-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/.env --project-name roottestserverkeepalive-gw8 --file /ClickHouse/tests/integration/test_server_keep_alive/_instances-0-gw8/node/docker-compose.yml down --volumes] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_s3_table_functions/configs/config.d/minio.xml'] to /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_s3_table_functions/configs/config.d/minio.xml'] to /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/database Setup database dir /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/database Setup logs dir /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/logs Setup logs dir /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] Executing query SELECT * FROM s3_test order by id FORMAT Values on node Executing query SELECT * FROM s3_test order by id FORMAT Values on node run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:10 Executing query DROP TABLE IF EXISTS s3_test SYNC on node Executing query DROP TABLE IF EXISTS s3_test SYNC on node [gw1] PASSED test_s3_with_https/test.py::test_s3_with_https[s3_secure] test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy] Executing query CREATE TABLE s3_test ( id Int64, data String ) ENGINE=MergeTree() ORDER BY id SETTINGS storage_policy='s3_secure_with_proxy' on node Executing query CREATE TABLE s3_test ( id Int64, data String ) ENGINE=MergeTree() ORDER BY id SETTINGS storage_policy='s3_secure_with_proxy' on node Executing query INSERT INTO s3_test VALUES (0,'data'),(1,'data') on node Executing query INSERT INTO s3_test VALUES (0,'data'),(1,'data') on node Executing query SELECT * FROM s3_test order by id FORMAT Values on node Executing query SELECT * FROM s3_test order by id FORMAT Values on node Executing query DROP TABLE IF EXISTS s3_test SYNC on node Executing query DROP TABLE IF EXISTS s3_test SYNC on node Stderr: Container roottestserverkeepalive-gw8-node-1 Stopping Stderr: Container roottestserverkeepalive-gw8-node-1 Stopping Stderr: Container roottestserverkeepalive-gw8-node-1 Stopped Stderr: Container roottestserverkeepalive-gw8-node-1 Stopped Stderr: Container roottestserverkeepalive-gw8-node-1 Removing Stderr: Container roottestserverkeepalive-gw8-node-1 Removing Stderr: Container roottestserverkeepalive-gw8-node-1 Removed Stderr: Container roottestserverkeepalive-gw8-node-1 Removed Stderr: Network roottestserverkeepalive-gw8_default Removing Stderr: Network roottestserverkeepalive-gw8_default Removing Stderr: Network roottestserverkeepalive-gw8_default Removed Stderr: Network roottestserverkeepalive-gw8_default Removed Cleanup called Cleanup called Docker networks for project roottestserverkeepalive-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestserverkeepalive-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestserverkeepalive-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestserverkeepalive-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-proxy1-1/logs?stderr=1&stdout=1×tamps=0&follow=0&tail=all HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-proxy1-1/logs?stderr=1&stdout=1×tamps=0&follow=0&tail=all HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-proxy1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3withhttps-gw1-proxy1-1/json HTTP/1.1" 200 None Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] [gw1] PASSED test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy] Docker volumes for project roottestserverkeepalive-gw8 are DRIVER VOLUME NAME Docker volumes for project roottestserverkeepalive-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestserverkeepalive-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestserverkeepalive-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestserverkeepalive-gw8 No running containers for project: roottestserverkeepalive-gw8 Trying to prune unused networks... Trying to prune unused networks... run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 test_storage_url_with_proxy/test.py::test_s3_with_proxy_list ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 5adb7eee8142 ENV HOSTNAME 5adb7eee8142 ENV SHLVL 0 ENV SHLVL 0 ENV HOME /root ENV HOME /root ENV OLDPWD / ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30400 30401 30402 30403 30404 30405 30406 30407 30408 30409 30410 30411 30412 30413 30414 30415 30416 30417 30418 30419 30420 30421 30422 30423 30424 30425 30426 30427 30428 30429 30430 30431 30432 30433 30434 30435 30436 30437 30438 30439 30440 30441 30442 30443 30444 30445 30446 30447 30448 30449 ENV WORKER_FREE_PORTS 30400 30401 30402 30403 30404 30405 30406 30407 30408 30409 30410 30411 30412 30413 30414 30415 30416 30417 30418 30419 30420 30421 30422 30423 30424 30425 30426 30427 30428 30429 30430 30431 30432 30433 30434 30435 30436 30437 30438 30439 30440 30441 30442 30443 30444 30445 30446 30447 30448 30449 ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_WORKER gw8 ENV PYTEST_XDIST_WORKER gw8 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_storage_url_with_proxy/test.py::test_s3_with_proxy_list (setup) ENV PYTEST_CURRENT_TEST test_storage_url_with_proxy/test.py::test_s3_with_proxy_list (setup) CLUSTER INIT base_config_dir:/clickhouse-config CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:rootteststorageurlwithproxy-gw8. Added instance name:proxy_list_node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env', '--project-name', 'rootteststorageurlwithproxy-gw8', '--file', '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:rootteststorageurlwithproxy-gw8. Added instance name:proxy_list_node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env', '--project-name', 'rootteststorageurlwithproxy-gw8', '--file', '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Starting cluster... Starting cluster... Running tests in /ClickHouse/tests/integration/test_storage_url_with_proxy/test.py Running tests in /ClickHouse/tests/integration/test_storage_url_with_proxy/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project rootteststorageurlwithproxy-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurlwithproxy-gw8 are NETWORK ID NAME DRIVER SCOPE run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestthrottling-gw7-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestthrottling-gw7-node-1/exec HTTP/1.1" 201 74 Docker containers for project rootteststorageurlwithproxy-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurlwithproxy-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststorageurlwithproxy-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurlwithproxy-gw8 are DRIVER VOLUME NAME Cleanup called Cleanup called http://localhost:None "POST /v1.46/exec/6fe7816d3bcedeab54c16beb14daa4c504eaed3ada6a6be196a4257ed070c7d1/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/6fe7816d3bcedeab54c16beb14daa4c504eaed3ada6a6be196a4257ed070c7d1/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/6fe7816d3bcedeab54c16beb14daa4c504eaed3ada6a6be196a4257ed070c7d1/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/6fe7816d3bcedeab54c16beb14daa4c504eaed3ada6a6be196a4257ed070c7d1/json HTTP/1.1" 200 586 Stdout:772 Stdout:772 Docker networks for project rootteststorageurlwithproxy-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurlwithproxy-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststorageurlwithproxy-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurlwithproxy-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststorageurlwithproxy-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurlwithproxy-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageurlwithproxy-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageurlwithproxy-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageurlwithproxy-gw8 No running containers for project: rootteststorageurlwithproxy-gw8 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 Setup directory for instance: proxy_list_node Setup directory for instance: proxy_list_node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_storage_url_with_proxy/configs/config.d/proxy_list.xml'] to /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_storage_url_with_proxy/configs/config.d/proxy_list.xml'] to /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/database Setup database dir /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/database Setup logs dir /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/logs Setup logs dir /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/f33fda4a6cdfecfde4fef7a6d89c6d0ba0cdbf2251583739634db54ea97fbd15/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/f33fda4a6cdfecfde4fef7a6d89c6d0ba0cdbf2251583739634db54ea97fbd15/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/f33fda4a6cdfecfde4fef7a6d89c6d0ba0cdbf2251583739634db54ea97fbd15/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/f33fda4a6cdfecfde4fef7a6d89c6d0ba0cdbf2251583739634db54ea97fbd15/json HTTP/1.1" 200 586 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2388 Stdout:2388 Clickhouse process running. Clickhouse process running. run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestthrottling-gw7-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestthrottling-gw7-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2388 Stdout:2388 Executing query select 20 on node Executing query select 20 on node No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/exec HTTP/1.1" 201 74 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/aaf824f9fda7908bebfff36862b9584d23ace2a62eee38b50fe38238dd0042c9/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/aaf824f9fda7908bebfff36862b9584d23ace2a62eee38b50fe38238dd0042c9/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/aaf824f9fda7908bebfff36862b9584d23ace2a62eee38b50fe38238dd0042c9/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/aaf824f9fda7908bebfff36862b9584d23ace2a62eee38b50fe38238dd0042c9/json HTTP/1.1" 200 586 Command:[docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:764 Stdout:764 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:764 Stdout:764 Executing query select 20 on node_default Executing query select 20 on node_default Executing query select 1 on node_default Executing query select 1 on node_default Executing query system flush logs on node_default Executing query system flush logs on node_default Executing query select 2 on node_default Executing query select 2 on node_default Executing query system flush logs on node_default Executing query system flush logs on node_default Executing query select count() >= 2 from system.query_log on node_default Executing query select count() >= 2 from system.query_log on node_default http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query drop table system.query_log sync on node_default Executing query drop table system.query_log sync on node_default Stdout:1564 Stdout:1564 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1564 Stdout:1564 Executing query select 20 on replica1 Executing query select 20 on replica1 Executing query select 3 on node_default Executing query select 3 on node_default Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_atomic' FORMAT Values on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_atomic' FORMAT Values on replica1 Executing query system flush logs on node_default Executing query system flush logs on node_default Executing query DROP DATABASE test_db_atomic on replica1 Executing query DROP DATABASE test_db_atomic on replica1 Executing query select count() >= 1 from system.query_log on node_default Executing query select count() >= 1 from system.query_log on node_default [gw4] PASSED test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table] test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table] Executing query CREATE DATABASE IF NOT EXISTS test_db_lazy ENGINE=Lazy(60); on replica1 Executing query CREATE DATABASE IF NOT EXISTS test_db_lazy ENGINE=Lazy(60); on replica1 Executing query drop table system.query_log sync on node_default Executing query drop table system.query_log sync on node_default Executing query CREATE TABLE test_db_lazy.test_table (n Int64) ENGINE=Log ; on replica1 Executing query CREATE TABLE test_db_lazy.test_table (n Int64) ENGINE=Log ; on replica1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Executing query CREATE TABLE test_db_lazy.test_table_perm (n Int64) ENGINE=Log ; on replica1 Executing query CREATE TABLE test_db_lazy.test_table_perm (n Int64) ENGINE=Log ; on replica1 Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 764 ? 00:00:00 clickhouse Stdout: 764 ? 00:00:00 clickhouse run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Executing query SELECT uuid FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table' on replica1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:764 Stdout:764 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT * FROM system.detached_tables WHERE database='test_db_lazy' on replica1 Executing query SELECT * FROM system.detached_tables WHERE database='test_db_lazy' on replica1 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query SET database_replicated_always_detach_permanently=1; DETACH TABLE test_db_lazy.test_table on replica1 Executing query SET database_replicated_always_detach_permanently=1; DETACH TABLE test_db_lazy.test_table on replica1 Executing query DETACH TABLE test_db_lazy.test_table_perm PERMANENTLY on replica1 Executing query DETACH TABLE test_db_lazy.test_table_perm PERMANENTLY on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_lazy' FORMAT Values on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_lazy' FORMAT Values on replica1 [gw4] PASSED test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table] test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table] Executing query CREATE DATABASE IF NOT EXISTS test_db_repl ENGINE=Replicated('/clickhouse/tables/test_table','shard1', 'replica1'); on replica1 Executing query CREATE DATABASE IF NOT EXISTS test_db_repl ENGINE=Replicated('/clickhouse/tables/test_table','shard1', 'replica1'); on replica1 Executing query CREATE TABLE test_db_repl.test_table (n Int64) ENGINE=ReplicatedMergeTree ORDER BY n; on replica1 Executing query CREATE TABLE test_db_repl.test_table (n Int64) ENGINE=ReplicatedMergeTree ORDER BY n; on replica1 Executing query CREATE TABLE test_db_repl.test_table_perm (n Int64) ENGINE=ReplicatedMergeTree ORDER BY n; on replica1 Executing query CREATE TABLE test_db_repl.test_table_perm (n Int64) ENGINE=ReplicatedMergeTree ORDER BY n; on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT uuid FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table_perm' on replica1 Executing query SELECT metadata_path FROM system.tables WHERE table='test_table_perm' on replica1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT * FROM system.detached_tables WHERE database='test_db_repl' on replica1 Executing query SELECT * FROM system.detached_tables WHERE database='test_db_repl' on replica1 Stdout:764 Stdout:764 Executing query SET database_replicated_always_detach_permanently=1; DETACH TABLE test_db_repl.test_table on replica1 Executing query SET database_replicated_always_detach_permanently=1; DETACH TABLE test_db_repl.test_table on replica1 Executing query DETACH TABLE test_db_repl.test_table_perm PERMANENTLY on replica1 Executing query DETACH TABLE test_db_repl.test_table_perm PERMANENTLY on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_repl' FORMAT Values on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_repl' FORMAT Values on replica1 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 1564 ? 00:00:00 clickhouse Stdout: 1564 ? 00:00:00 clickhouse run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c pkill clickhouse] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1564 Stdout:1564 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:764 Stdout:764 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1564 Stdout:1564 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:764 Stdout:764 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 http://172.16.6.7:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix=data%2F HTTP/1.1" 200 0 Executing query DROP TABLE IF EXISTS s3_test SYNC on node1 Executing query DROP TABLE IF EXISTS s3_test SYNC on node1 Executing query DROP TABLE IF EXISTS s3_test SYNC on node2 Executing query DROP TABLE IF EXISTS s3_test SYNC on node2 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3] test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op] Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1564 Stdout:1564 Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ on node2 Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ VALUES (1, '2024-10-23') on node1 Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ VALUES (1, '2024-10-23') on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ on node2 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_' and active and disk_name='s31' on node2 Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_' and active and disk_name='s31' on node2 Connected to Mongo dbs: ['admin', 'config', 'local'] Connected to Mongo dbs: ['admin', 'config', 'local'] ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ on node2 http://localhost:None "POST /v1.46/exec/4eba7d76508f3410491e255a097371c37b47ca1c7daf6e2669b8b587e8e07162/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/4eba7d76508f3410491e255a097371c37b47ca1c7daf6e2669b8b587e8e07162/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/4eba7d76508f3410491e255a097371c37b47ca1c7daf6e2669b8b587e8e07162/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/4eba7d76508f3410491e255a097371c37b47ca1c7daf6e2669b8b587e8e07162/json HTTP/1.1" 200 586 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ SYNC on node2 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_after_op_ SYNC on node2 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] [gw0] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op] test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op] Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Stdout:1564 Stdout:1564 Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.5... Waiting for ClickHouse start in node, ip: 172.16.1.5... http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ on node2 http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ VALUES (1, '2024-10-23') on node1 Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ VALUES (1, '2024-10-23') on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/904c4c94d019605bc26a0bbd350bd12f7e8a5f0cb49da5c12b2c3200d702a7ab/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ on node2 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_' and active and disk_name='s31' on node2 Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_' and active and disk_name='s31' on node2 Stdout:1563 Stdout:1563 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ on node2 Stdout:1563 Stdout:1563 Executing query select 20 on node_default Executing query select 20 on node_default Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ SYNC on node1 Executing query system flush logs on node_default Executing query system flush logs on node_default Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ SYNC on node2 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_None_zero_copy_lock_zk_fail_before_op_ SYNC on node2 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select count() >= 0 from system.query_log on node_default Executing query select count() >= 0 from system.query_log on node_default run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] [gw0] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op] test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op] Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/exec HTTP/1.1" 201 74 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/yyy-override-query_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/yyy-override-query_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/yyy-override-query_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/yyy-override-query_log.xml] http://localhost:None "POST /v1.46/exec/f17805a6537e492ead8ade7d8d3bc08d3f02932cb90f2a1c54abb5f6ba2d0f8e/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/f17805a6537e492ead8ade7d8d3bc08d3f02932cb90f2a1c54abb5f6ba2d0f8e/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/f17805a6537e492ead8ade7d8d3bc08d3f02932cb90f2a1c54abb5f6ba2d0f8e/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/f17805a6537e492ead8ade7d8d3bc08d3f02932cb90f2a1c54abb5f6ba2d0f8e/json HTTP/1.1" 200 586 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ on node2 Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 1563 ? 00:00:01 clickhouse Stdout: 1563 ? 00:00:01 clickhouse run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ VALUES (1, '2024-10-23') on node1 Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ VALUES (1, '2024-10-23') on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1563 Stdout:1563 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_unlock_zk_fail_after_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_unlock_zk_fail_after_op on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SELECT COUNT() FROM mongodb('mongo_secure:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String', options='ssl=true') on node Executing query SELECT COUNT() FROM mongodb('mongo_secure:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String', options='ssl=true') on node Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_unlock_zk_fail_after_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_unlock_zk_fail_after_op on node1 Executing query SELECT sum(key) FROM mongodb('mongo_secure:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String', options='ssl=true') on node Executing query SELECT sum(key) FROM mongodb('mongo_secure:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String', options='ssl=true') on node Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query SELECT sum(key) FROM mongodb('mongo_secure:27017', 'test', 'simple_table', 'root', 'clickhouse', 'key UInt64, data String', 'ssl=true') on node Executing query SELECT sum(key) FROM mongodb('mongo_secure:27017', 'test', 'simple_table', 'root', 'clickhouse', 'key UInt64, data String', 'ssl=true') on node Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SELECT data from mongodb('mongo_secure:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String', options='ssl=true') where key = 42 on node Executing query SELECT data from mongodb('mongo_secure:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String', options='ssl=true') where key = 42 on node Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_' and active and disk_name='s31' on node2 Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_' and active and disk_name='s31' on node2 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ on node2 [gw3] PASSED test_table_function_mongodb_legacy/test.py::test_secure_connection[True] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml stop --timeout 20] test_table_function_mongodb_legacy/test.py::test_simple_select[False] Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ SYNC on node2 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_after_op_ SYNC on node2 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2359 Stdout:2359 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] [gw0] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op] test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op] Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Stdout:2359 Stdout:2359 Executing query select 20 on replica1 Executing query select 20 on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_repl' FORMAT Values on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_repl' FORMAT Values on replica1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ on node2 Stdout:1563 Stdout:1563 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 2359 ? 00:00:00 clickhouse Stdout: 2359 ? 00:00:00 clickhouse run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemdetachedtables-gw4-replica1-1 bash -c pkill clickhouse] Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ VALUES (1, '2024-10-23') on node1 Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ VALUES (1, '2024-10-23') on node1 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2359 Stdout:2359 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: resolver Pulling Stderr: resolver Pulling Stderr: proxy1 Pulling Stderr: proxy1 Pulling Stderr: minio1 Pulling Stderr: minio1 Pulling Stderr: proxy_list_node Pulling Stderr: proxy_list_node Pulling Stderr: resolver Pulled Stderr: resolver Pulled Stderr: proxy1 Pulled Stderr: proxy1 Pulled Stderr: proxy_list_node Pulled Stderr: proxy_list_node Pulled Stderr: minio1 Pulled Stderr: minio1 Pulled Trying to create Minio instance by command docker compose --project-name rootteststorageurlwithproxy-gw8 --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Trying to create Minio instance by command docker compose --project-name rootteststorageurlwithproxy-gw8 --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Command:[docker compose --project-name rootteststorageurlwithproxy-gw8 --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Command:[docker compose --project-name rootteststorageurlwithproxy-gw8 --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: minio1 Pulling Stderr: minio1 Pulling Stderr: proxy1 Pulling Stderr: proxy1 Pulling Stderr: node Pulling Stderr: node Pulling Stderr: resolver Pulling Stderr: resolver Pulling Stderr: resolver Pulled Stderr: resolver Pulled Stderr: minio1 Pulled Stderr: minio1 Pulled Stderr: proxy1 Pulled Stderr: proxy1 Pulled Stderr: node Pulled Stderr: node Pulled Trying to create Minio instance by command docker compose --project-name roottests3tablefunctions-gw5 --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Trying to create Minio instance by command docker compose --project-name roottests3tablefunctions-gw5 --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Command:[docker compose --project-name roottests3tablefunctions-gw5 --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Command:[docker compose --project-name roottests3tablefunctions-gw5 --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_unlock_zk_fail_after_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_unlock_zk_fail_after_op on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM DISABLE FAILPOINT zero_copy_unlock_zk_fail_after_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_unlock_zk_fail_after_op on node1 Stdout:1563 Stdout:1563 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2359 Stdout:2359 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Stderr:time="2025-04-02T04:15:06Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:15:06Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network rootteststorageurlwithproxy-gw8_default Creating Stderr: Network rootteststorageurlwithproxy-gw8_default Creating Stderr: Network rootteststorageurlwithproxy-gw8_default Created Stderr: Network rootteststorageurlwithproxy-gw8_default Created Stderr: Volume "rootteststorageurlwithproxy-gw8_data1-1" Creating Stderr: Volume "rootteststorageurlwithproxy-gw8_data1-1" Creating Stderr: Volume "rootteststorageurlwithproxy-gw8_data1-1" Created Stderr: Volume "rootteststorageurlwithproxy-gw8_data1-1" Created Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Started Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Started Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Started Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Started Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Started Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Started Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Started Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Started Stderr:time="2025-04-02T04:15:07Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:07Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:07Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:07Z" level=debug msg="otel error" error="" Trying to connect to Minio... Trying to connect to Minio... get_instance_ip instance_name=minio1 get_instance_ip instance_name=minio1 http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-minio1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-minio1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=proxy1 get_instance_ip instance_name=proxy1 http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy1-1/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.3.4:9001 Starting new HTTP connection (1): 172.16.3.4:9001 Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (2): 172.16.3.4:9001 Starting new HTTP connection (2): 172.16.3.4:9001 Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (3): 172.16.3.4:9001 Starting new HTTP connection (3): 172.16.3.4:9001 Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (4): 172.16.3.4:9001 Starting new HTTP connection (4): 172.16.3.4:9001 Can't connect to Minio: HTTPConnectionPool(host='172.16.3.4', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Can't connect to Minio: HTTPConnectionPool(host='172.16.3.4', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_' and active and disk_name='s31' on node2 Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_' and active and disk_name='s31' on node2 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ on node2 Stderr:time="2025-04-02T04:15:06Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:15:06Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottests3tablefunctions-gw5_default Creating Stderr: Network roottests3tablefunctions-gw5_default Creating Stderr: Network roottests3tablefunctions-gw5_default Created Stderr: Network roottests3tablefunctions-gw5_default Created Stderr: Volume "roottests3tablefunctions-gw5_data1-1" Creating Stderr: Volume "roottests3tablefunctions-gw5_data1-1" Creating Stderr: Volume "roottests3tablefunctions-gw5_data1-1" Created Stderr: Volume "roottests3tablefunctions-gw5_data1-1" Created Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Creating Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Creating Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Creating Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Creating Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Created Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Created Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Created Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Created Stderr: Container roottests3tablefunctions-gw5-minio1-1 Creating Stderr: Container roottests3tablefunctions-gw5-minio1-1 Creating Stderr: Container roottests3tablefunctions-gw5-resolver-1 Creating Stderr: Container roottests3tablefunctions-gw5-resolver-1 Creating Stderr: Container roottests3tablefunctions-gw5-resolver-1 Created Stderr: Container roottests3tablefunctions-gw5-resolver-1 Created Stderr: Container roottests3tablefunctions-gw5-minio1-1 Created Stderr: Container roottests3tablefunctions-gw5-minio1-1 Created Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Starting Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Starting Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Starting Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Starting Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Started Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Started Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Started Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Started Stderr: Container roottests3tablefunctions-gw5-resolver-1 Starting Stderr: Container roottests3tablefunctions-gw5-resolver-1 Starting Stderr: Container roottests3tablefunctions-gw5-minio1-1 Starting Stderr: Container roottests3tablefunctions-gw5-minio1-1 Starting Stderr: Container roottests3tablefunctions-gw5-resolver-1 Started Stderr: Container roottests3tablefunctions-gw5-resolver-1 Started Stderr: Container roottests3tablefunctions-gw5-minio1-1 Started Stderr: Container roottests3tablefunctions-gw5-minio1-1 Started Stderr:time="2025-04-02T04:15:07Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:07Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:07Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:07Z" level=debug msg="otel error" error="" Trying to connect to Minio... Trying to connect to Minio... get_instance_ip instance_name=minio1 get_instance_ip instance_name=minio1 http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-minio1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-minio1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=proxy1 get_instance_ip instance_name=proxy1 http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-proxy1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-proxy1-1/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.5.5:9001 Starting new HTTP connection (1): 172.16.5.5:9001 Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (2): 172.16.5.5:9001 Starting new HTTP connection (2): 172.16.5.5:9001 Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (3): 172.16.5.5:9001 Starting new HTTP connection (3): 172.16.5.5:9001 Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (4): 172.16.5.5:9001 Starting new HTTP connection (4): 172.16.5.5:9001 Can't connect to Minio: HTTPConnectionPool(host='172.16.5.5', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Can't connect to Minio: HTTPConnectionPool(host='172.16.5.5', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ SYNC on node2 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_after_op_zero_copy_lock_zk_fail_before_op_ SYNC on node2 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] [gw0] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op] test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op] Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 No clickhouse process running. Start new one. No clickhouse process running. Start new one. run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/621f4c12f7f6f68ff7de59414a17e66ea76cf59d1a2caa86f70b5856593942ca/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/621f4c12f7f6f68ff7de59414a17e66ea76cf59d1a2caa86f70b5856593942ca/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/621f4c12f7f6f68ff7de59414a17e66ea76cf59d1a2caa86f70b5856593942ca/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/621f4c12f7f6f68ff7de59414a17e66ea76cf59d1a2caa86f70b5856593942ca/json HTTP/1.1" 200 586 Stdout:2359 Stdout:2359 Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ VALUES (1, '2024-10-23') on node1 Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ VALUES (1, '2024-10-23') on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_unlock_zk_fail_before_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_unlock_zk_fail_before_op on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Starting new HTTP connection (5): 172.16.3.4:9001 Starting new HTTP connection (5): 172.16.3.4:9001 http://172.16.3.4:9001 "GET / HTTP/1.1" 200 0 http://172.16.3.4:9001 "GET / HTTP/1.1" 200 0 Connected to Minio. Connected to Minio. http://172.16.3.4:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.3.4:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.3.4:9001 "PUT /root HTTP/1.1" 200 0 http://172.16.3.4:9001 "PUT /root HTTP/1.1" 200 0 S3 bucket 'root' created S3 bucket 'root' created http://172.16.3.4:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.3.4:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.3.4:9001 "PUT /root2 HTTP/1.1" 200 0 http://172.16.3.4:9001 "PUT /root2 HTTP/1.1" 200 0 S3 bucket 'root2' created S3 bucket 'root2' created ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_after_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_unlock_zk_fail_before_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_unlock_zk_fail_before_op on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Starting new HTTP connection (5): 172.16.5.5:9001 Starting new HTTP connection (5): 172.16.5.5:9001 http://172.16.5.5:9001 "GET / HTTP/1.1" 200 0 http://172.16.5.5:9001 "GET / HTTP/1.1" 200 0 Connected to Minio. Connected to Minio. http://172.16.5.5:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.5.5:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.5.5:9001 "PUT /root HTTP/1.1" 200 0 http://172.16.5.5:9001 "PUT /root HTTP/1.1" 200 0 S3 bucket 'root' created S3 bucket 'root' created http://172.16.5.5:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.5.5:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.5.5:9001 "PUT /root2 HTTP/1.1" 200 0 http://172.16.5.5:9001 "PUT /root2 HTTP/1.1" 200 0 S3 bucket 'root2' created S3 bucket 'root2' created ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_' GROUP BY disk_name on node1 Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Running Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Running Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Running Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Running Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Running Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Running Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Running Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Running Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Creating Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Created Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Starting Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Started Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=proxy_list_node get_instance_ip instance_name=proxy_list_node http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy_list_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy_list_node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=proxy_list_node get_instance_ip instance_name=proxy_list_node http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy_list_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy_list_node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in proxy_list_node, ip: 172.16.3.6... Waiting for ClickHouse start in proxy_list_node, ip: 172.16.3.6... http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy_list_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy_list_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_' and active and disk_name='s31' on node2 Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_' and active and disk_name='s31' on node2 http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ on node2 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ on node2 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Running Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Running Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Running Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Running Stderr: Container roottests3tablefunctions-gw5-minio1-1 Running Stderr: Container roottests3tablefunctions-gw5-minio1-1 Running Stderr: Container roottests3tablefunctions-gw5-resolver-1 Running Stderr: Container roottests3tablefunctions-gw5-resolver-1 Running Stderr: Container roottests3tablefunctions-gw5-node-1 Creating Stderr: Container roottests3tablefunctions-gw5-node-1 Creating Stderr: Container roottests3tablefunctions-gw5-node-1 Created Stderr: Container roottests3tablefunctions-gw5-node-1 Created Stderr: Container roottests3tablefunctions-gw5-node-1 Starting Stderr: Container roottests3tablefunctions-gw5-node-1 Starting Stderr: Container roottests3tablefunctions-gw5-node-1 Started Stderr: Container roottests3tablefunctions-gw5-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.5.6... Waiting for ClickHouse start in node, ip: 172.16.5.6... http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottests3tablefunctions-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None Stdout:2358 Stdout:2358 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2359 Stdout:2359 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ SYNC on node1 Stdout:2358 Stdout:2358 Executing query select 20 on node_default Executing query select 20 on node_default http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ SYNC on node2 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_after_op_ SYNC on node2 http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None [gw9] PASSED test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate Executing query SYSTEM FLUSH LOGS on node_default Executing query SYSTEM FLUSH LOGS on node_default http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/07be6fd7736d46eb5cf0ca43a9c136c4c6a01faed22850c95927a0ce4275ff55/json HTTP/1.1" 200 None ClickHouse proxy_list_node started ClickHouse proxy_list_node started Cluster started Cluster started Executing query INSERT INTO FUNCTION s3('http://minio1:9001/root/data/ch-proxy-test/test.csv', 'minio', 'minio123', 'CSV', 'key String, value String') SETTINGS s3_truncate_on_insert=1 VALUES ('color','red'),('size','10') on proxy_list_node Executing query INSERT INTO FUNCTION s3('http://minio1:9001/root/data/ch-proxy-test/test.csv', 'minio', 'minio123', 'CSV', 'key String, value String') SETTINGS s3_truncate_on_insert=1 VALUES ('color','red'),('size','10') on proxy_list_node http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None [gw0] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op] test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op] Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Executing query CREATE TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ ON CLUSTER test_cluster (num UInt64, date DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_', '{replica}') ORDER BY date PARTITION BY date SETTINGS storage_policy='hybrid' on node1 Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SHOW CREATE TABLE system.query_log on node_default http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SELECT * FROM url('http://minio1:9001/root/data/ch-proxy-test/test.csv', 'CSV', 'a String, b String',headers('Host'='minio1', 'Date'= 'Wed, 02 Apr 2025 04:15:09 +0000', 'Content-Type'='application/zstd','Authorization'='AWS minio:RA4bL7Hh4B50HEZXLp+B6BZ5VEA=')) FORMAT Values on proxy_list_node Executing query SELECT * FROM url('http://minio1:9001/root/data/ch-proxy-test/test.csv', 'CSV', 'a String, b String',headers('Host'='minio1', 'Date'= 'Wed, 02 Apr 2025 04:15:09 +0000', 'Content-Type'='application/zstd','Authorization'='AWS minio:RA4bL7Hh4B50HEZXLp+B6BZ5VEA=')) FORMAT Values on proxy_list_node Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bc1047975720e051d36c41f7e551772aee5f58b3724a1388f2ca7af76b0fbcff/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Cluster started Cluster started Executing query INSERT INTO FUNCTION s3 ( nc_s3, filename = 'test_file.tsv.gz', format = 'TSV', structure = 'number UInt64', compression_method = 'gz' ) SELECT * FROM numbers(1000000) on node Executing query INSERT INTO FUNCTION s3 ( nc_s3, filename = 'test_file.tsv.gz', format = 'TSV', structure = 'number UInt64', compression_method = 'gz' ) SELECT * FROM numbers(1000000) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml down --volumes] Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SYSTEM STOP FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ VALUES (1, '2024-10-23') on node1 Executing query INSERT INTO table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ VALUES (1, '2024-10-23') on node1 get_instance_ip instance_name=minio1 get_instance_ip instance_name=minio1 http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-minio1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-minio1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy1-1/logs?stderr=1&stdout=1×tamps=0&follow=0&tail=all HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy1-1/logs?stderr=1&stdout=1×tamps=0&follow=0&tail=all HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageurlwithproxy-gw8-proxy1-1/json HTTP/1.1" 200 None Method with ip: GET http://172.16.3.4 Method with ip: GET http://172.16.3.4 Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] [gw8] PASSED test_storage_url_with_proxy/test.py::test_s3_with_proxy_list Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_unlock_zk_fail_before_op on node1 Executing query SYSTEM ENABLE FAILPOINT zero_copy_unlock_zk_fail_before_op on node1 Executing query SELECT count(*) FROM s3 ( nc_s3, filename = 'test_file.tsv.gz', format = 'TSV', structure = 'number UInt64', compression_method = 'gz' ); on node Executing query SELECT count(*) FROM s3 ( nc_s3, filename = 'test_file.tsv.gz', format = 'TSV', structure = 'number UInt64', compression_method = 'gz' ); on node Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query SHOW CREATE TABLE system.query_thread_log on node_default Executing query SHOW CREATE TABLE system.query_thread_log on node_default [gw5] PASSED test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Executing query SHOW CREATE TABLE system.query_thread_log on node_default Executing query SHOW CREATE TABLE system.query_thread_log on node_default http://localhost:None "POST /v1.46/containers/create HTTP/1.1" 201 88 http://localhost:None "POST /v1.46/containers/create HTTP/1.1" 201 88 http://localhost:None "GET /v1.46/containers/61c46c4c77dc4e6ccf043ae70c45fdff65a349c86b7c871b655739619d1d82c5/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/61c46c4c77dc4e6ccf043ae70c45fdff65a349c86b7c871b655739619d1d82c5/json HTTP/1.1" 200 None Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_lock_zk_fail_before_op on node1 http://localhost:None "POST /v1.46/containers/61c46c4c77dc4e6ccf043ae70c45fdff65a349c86b7c871b655739619d1d82c5/start HTTP/1.1" 204 0 http://localhost:None "POST /v1.46/containers/61c46c4c77dc4e6ccf043ae70c45fdff65a349c86b7c871b655739619d1d82c5/start HTTP/1.1" 204 0 [network] Created new container 61c46c4c77dc4e6ccf043ae70c45fdff65a349c86b7c871b655739619d1d82c5 [network] Created new container 61c46c4c77dc4e6ccf043ae70c45fdff65a349c86b7c871b655739619d1d82c5 run container_id:roottests3tablefunctions-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'tc qdisc add dev eth0 root netem delay 1200ms'] run container_id:roottests3tablefunctions-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'tc qdisc add dev eth0 root netem delay 1200ms'] Command:[docker exec -u root roottests3tablefunctions-gw5-node-1 bash -c tc qdisc add dev eth0 root netem delay 1200ms] Command:[docker exec -u root roottests3tablefunctions-gw5-node-1 bash -c tc qdisc add dev eth0 root netem delay 1200ms] Executing query SHOW CREATE TABLE system.part_log on node_default Executing query SHOW CREATE TABLE system.part_log on node_default Executing query INSERT INTO FUNCTION s3 ( nc_s3, filename = 'test_file.tsv.gz', format = 'TSV', structure = 'number UInt64', compression_method = 'gz' ) SELECT * FROM numbers(1000000) on node Executing query INSERT INTO FUNCTION s3 ( nc_s3, filename = 'test_file.tsv.gz', format = 'TSV', structure = 'number UInt64', compression_method = 'gz' ) SELECT * FROM numbers(1000000) on node Executing query SYSTEM DISABLE FAILPOINT zero_copy_unlock_zk_fail_before_op on node1 Executing query SYSTEM DISABLE FAILPOINT zero_copy_unlock_zk_fail_before_op on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 Executing query ALTER TABLE table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ MOVE PARTITION '2024-10-23' TO VOLUME 'external' on node1 run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SHOW CREATE TABLE system.part_log on node_default Executing query SHOW CREATE TABLE system.part_log on node_default Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removed Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removing Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removing Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removed Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removed Cleanup called Cleanup called run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Executing query SELECT disk_name FROM system.parts WHERE table='table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_' GROUP BY disk_name on node1 Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemdetachedtables-gw4-replica1-1/exec HTTP/1.1" 201 74 Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] http://localhost:None "POST /v1.46/exec/9418d8fec5d749c6069a86e931be08a1bfbb568149dda4dc71fbea0aa06ab05b/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/9418d8fec5d749c6069a86e931be08a1bfbb568149dda4dc71fbea0aa06ab05b/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/9418d8fec5d749c6069a86e931be08a1bfbb568149dda4dc71fbea0aa06ab05b/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/9418d8fec5d749c6069a86e931be08a1bfbb568149dda4dc71fbea0aa06ab05b/json HTTP/1.1" 200 586 Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttablefunctionmongodblegacy-gw3 No running containers for project: roottesttablefunctionmongodblegacy-gw3 Trying to prune unused networks... Trying to prune unused networks... Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SYSTEM START FETCHES table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ on node2 Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:11 Stdout:11 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_' and active and disk_name='s31' on node2 Executing query select count() from system.parts where table = 'table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_' and active and disk_name='s31' on node2 Volumes pruned: 11 Volumes pruned: 11 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 5adb7eee8142 ENV HOSTNAME 5adb7eee8142 ENV SHLVL 0 ENV SHLVL 0 ENV HOME /root ENV HOME /root ENV OLDPWD / ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference test_s3_cluster/test.py::test_select_all test_s3_cluster/test.py::test_skip_unavailable_shards test_s3_cluster/test.py::test_union_all test_s3_cluster/test.py::test_unset_skip_unavailable_shards test_s3_cluster/test.py::test_wrong_cluster test_s3_table_functions/test.py::test_s3_table_functions test_s3_table_functions/test.py::test_s3_table_functions_timeouts 'test_s3_with_https/test.py::test_s3_with_https[s3_secure]' 'test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op]' 'test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3]' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10]' 'test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3]' test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side test_ssh_keys_authentication/test.py::test_ecdsa test_ssh_keys_authentication/test.py::test_ed25519 test_ssh_keys_authentication/test.py::test_key_with_passphrase test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase test_ssh_keys_authentication/test.py::test_rsa test_ssh_keys_authentication/test.py::test_wrong_key test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards test_storage_redis/test.py::test_create_table test_storage_redis/test.py::test_delete test_storage_redis/test.py::test_select_int test_storage_redis/test.py::test_simple_insert test_storage_redis/test.py::test_simple_select test_storage_redis/test.py::test_truncate test_storage_redis/test.py::test_update 'test_storage_url/test.py::test_file_formats[CSV]' 'test_storage_url/test.py::test_file_formats[JSONEachRow]' 'test_storage_url/test.py::test_file_formats[Parquet]' 'test_storage_url/test.py::test_file_formats[TSV]' test_storage_url/test.py::test_partition_by test_storage_url/test.py::test_table_function_url_access_rights test_storage_url/test.py::test_url_cluster test_storage_url/test.py::test_url_cluster_with_named_collection test_storage_url_http_headers/test.py::test_storage_url_http_headers test_storage_url_http_headers/test.py::test_storage_url_redirected_headers test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers test_storage_url_with_proxy/test.py::test_s3_with_proxy_list test_store_cleanup/test.py::test_store_cleanup test_structured_logging_json/test.py::test_structured_logging_json_format test_system_clusters_actual_information/test.py::test test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 'test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table]' 'test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table]' test_system_logs_comment/test.py::test_system_logs_comment test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated test_system_logs_recreate/test.py::test_drop_system_log test_system_logs_recreate/test.py::test_system_logs_recreate 'test_system_merges/test.py::test_merge_simple[]' 'test_system_merges/test.py::test_merge_simple[replicated]' 'test_system_merges/test.py::test_mutation_simple[]' 'test_system_merges/test.py::test_mutation_simple[replicated]' test_system_queries/test.py::test_DROP_DNS_CACHE test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY test_system_queries/test.py::test_system_flush_logs test_system_start_stop_listen/test.py::test_all_protocols test_system_start_stop_listen/test.py::test_custom_protocols test_system_start_stop_listen/test.py::test_default_protocols test_system_start_stop_listen/test.py::test_except 'test_table_function_mongodb_legacy/test.py::test_auth_source[False]' 'test_table_function_mongodb_legacy/test.py::test_complex_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False]' 'test_table_function_mongodb_legacy/test.py::test_missing_columns[False]' 'test_table_function_mongodb_legacy/test.py::test_no_credentials[False]' 'test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False]' 'test_table_function_mongodb_legacy/test.py::test_secure_connection[True]' 'test_table_function_mongodb_legacy/test.py::test_simple_select[False]' test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 'test_throttling/test.py::test_backup_throttling[no_local_throttling]' 'test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling]' 'test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30150 30151 30152 30153 30154 30155 30156 30157 30158 30159 30160 30161 30162 30163 30164 30165 30166 30167 30168 30169 30170 30171 30172 30173 30174 30175 30176 30177 30178 30179 30180 30181 30182 30183 30184 30185 30186 30187 30188 30189 30190 30191 30192 30193 30194 30195 30196 30197 30198 30199 ENV WORKER_FREE_PORTS 30150 30151 30152 30153 30154 30155 30156 30157 30158 30159 30160 30161 30162 30163 30164 30165 30166 30167 30168 30169 30170 30171 30172 30173 30174 30175 30176 30177 30178 30179 30180 30181 30182 30183 30184 30185 30186 30187 30188 30189 30190 30191 30192 30193 30194 30195 30196 30197 30198 30199 ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_TESTRUNUID 8c1a69d4df4d42e1a238d83ede5df9ca ENV PYTEST_XDIST_WORKER gw3 ENV PYTEST_XDIST_WORKER gw3 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_table_function_mongodb_legacy/test.py::test_simple_select[False] (setup) ENV PYTEST_CURRENT_TEST test_table_function_mongodb_legacy/test.py::test_simple_select[False] (setup) CLUSTER INIT base_config_dir:/clickhouse-config CLUSTER INIT base_config_dir:/clickhouse-config Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SHOW CREATE TABLE system.trace_log on node_default Removed :/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3 Removed :/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottesttablefunctionmongodblegacy-gw3. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env', '--project-name', 'roottesttablefunctionmongodblegacy-gw3', '--file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Cluster name: project_name:roottesttablefunctionmongodblegacy-gw3. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env', '--project-name', 'roottesttablefunctionmongodblegacy-gw3', '--file', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Running tests in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py Running tests in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Cleanup called Cleanup called Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SELECT sum(num) FROM table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ on node2 Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SHOW CREATE TABLE system.trace_log on node_default Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ SYNC on node1 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ SYNC on node1 Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttablefunctionmongodblegacy-gw3 No running containers for project: roottesttablefunctionmongodblegacy-gw3 Trying to prune unused networks... Trying to prune unused networks... Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:11 Stdout:11 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ SYNC on node2 Executing query DROP TABLE IF EXISTS table_test_move_shared_zero_copy_lock_fail_zero_copy_unlock_zk_fail_before_op_zero_copy_lock_zk_fail_before_op_ SYNC on node2 Executing query SHOW CREATE TABLE system.metric_log on node_default Executing query SHOW CREATE TABLE system.metric_log on node_default Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 11 Volumes pruned: 11 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_table_function_mongodb_legacy/mongo_secure_config/config.d/ssl_conf.xml', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/configs/feature_flag.xml'] to /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_table_function_mongodb_legacy/mongo_secure_config/config.d/ssl_conf.xml', '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/configs/feature_flag.xml'] to /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/database Setup database dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/database Setup logs dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs Setup logs dir /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MONGO_HOST': 'mongo1', 'MONGO_EXTERNAL_PORT': '30158', 'MONGO_INTERNAL_PORT': '27017', 'MONGO_NO_CRED_EXTERNAL_PORT': '30159', 'MONGO_NO_CRED_INTERNAL_PORT': '27017', 'MONGO_SECURE_EXTERNAL_PORT': '30160', 'MONGO_SECURE_INTERNAL_PORT': '27017', 'MONGO_SECURE_CONFIG_DIR': '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/mongo_secure_config'} stored in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MONGO_HOST': 'mongo1', 'MONGO_EXTERNAL_PORT': '30158', 'MONGO_INTERNAL_PORT': '27017', 'MONGO_NO_CRED_EXTERNAL_PORT': '30159', 'MONGO_NO_CRED_INTERNAL_PORT': '27017', 'MONGO_SECURE_EXTERNAL_PORT': '30160', 'MONGO_SECURE_INTERNAL_PORT': '27017', 'MONGO_SECURE_CONFIG_DIR': '/ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/mongo_secure_config'} stored in /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml pull] Executing query SHOW CREATE TABLE system.metric_log on node_default Executing query SHOW CREATE TABLE system.metric_log on node_default [gw0] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op] test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge Executing query DROP TABLE IF EXISTS concurrent_merge SYNC on node1 Executing query DROP TABLE IF EXISTS concurrent_merge SYNC on node1 Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Executing query DROP TABLE IF EXISTS concurrent_merge SYNC on node2 Executing query DROP TABLE IF EXISTS concurrent_merge SYNC on node2 Executing query SHOW CREATE TABLE system.error_log on node_default Executing query SHOW CREATE TABLE system.error_log on node_default Executing query CREATE TABLE concurrent_merge (id UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/concurrent_merge', '{replica}') ORDER BY id SETTINGS index_granularity=2, storage_policy='s3', remote_fs_execute_merges_on_single_replica_time_threshold=1 on node1 Executing query CREATE TABLE concurrent_merge (id UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/concurrent_merge', '{replica}') ORDER BY id SETTINGS index_granularity=2, storage_policy='s3', remote_fs_execute_merges_on_single_replica_time_threshold=1 on node1 Executing query SHOW CREATE TABLE system.error_log on node_default Executing query SHOW CREATE TABLE system.error_log on node_default Stderr: node_2 Skipped - Image is already being pulled by node_1 Stderr: node_2 Skipped - Image is already being pulled by node_1 Stderr: node_0 Skipped - Image is already being pulled by node_1 Stderr: node_0 Skipped - Image is already being pulled by node_1 Stderr: node_1 Pulling Stderr: node_1 Pulling Stderr: azurite1 Pulling Stderr: azurite1 Pulling Stderr: node_1 Pulled Stderr: node_1 Pulled Stderr: azurite1 Pulled Stderr: azurite1 Pulled Trying to create Azurite instance by command docker compose --project-name rootteststorageazureblobstoragecluster-gw2 --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d Trying to create Azurite instance by command docker compose --project-name rootteststorageazureblobstoragecluster-gw2 --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d Command:[docker compose --project-name rootteststorageazureblobstoragecluster-gw2 --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d] Command:[docker compose --project-name rootteststorageazureblobstoragecluster-gw2 --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d] Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query CREATE TABLE concurrent_merge (id UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/concurrent_merge', '{replica}') ORDER BY id SETTINGS index_granularity=2, storage_policy='s3', remote_fs_execute_merges_on_single_replica_time_threshold=1 on node2 Executing query CREATE TABLE concurrent_merge (id UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/concurrent_merge', '{replica}') ORDER BY id SETTINGS index_granularity=2, storage_policy='s3', remote_fs_execute_merges_on_single_replica_time_threshold=1 on node2 Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query system stop merges on node1 Executing query system stop merges on node1 Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query system stop merges on node2 Executing query system stop merges on node2 Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default Executing query insert into concurrent_merge select number from numbers(40) on node1 Executing query insert into concurrent_merge select number from numbers(40) on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-query_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-query_log.xml ] Executing query insert into concurrent_merge select number + 1 from numbers(40) on node1 Executing query insert into concurrent_merge select number + 1 from numbers(40) on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml ] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml ] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-part_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-part_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-part_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-part_log.xml ] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr:time="2025-04-02T04:15:11Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:15:11Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network rootteststorageazureblobstoragecluster-gw2_default Creating Stderr: Network rootteststorageazureblobstoragecluster-gw2_default Creating Stderr: Network rootteststorageazureblobstoragecluster-gw2_default Created Stderr: Network rootteststorageazureblobstoragecluster-gw2_default Created Stderr: Volume "rootteststorageazureblobstoragecluster-gw2_data1-1" Creating Stderr: Volume "rootteststorageazureblobstoragecluster-gw2_data1-1" Creating Stderr: Volume "rootteststorageazureblobstoragecluster-gw2_data1-1" Created Stderr: Volume "rootteststorageazureblobstoragecluster-gw2_data1-1" Created Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Creating Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Creating Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Created Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Created Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Starting Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Starting Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Started Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Started Stderr:time="2025-04-02T04:15:11Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:11Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:11Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:11Z" level=debug msg="otel error" error="" Trying to connect to Azurite Trying to connect to Azurite run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-trace_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-trace_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-trace_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-trace_log.xml ] Stdout:3153 Stdout:3153 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemdetachedtables-gw4-replica1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemdetachedtables-gw4-replica1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select count() from system.parts where table = 'concurrent_merge' and active on node2 Executing query select count() from system.parts where table = 'concurrent_merge' and active on node2 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-metric_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-metric_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-metric_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-metric_log.xml ] Stdout:3153 Stdout:3153 Executing query select 20 on replica1 Executing query select 20 on replica1 Executing query alter table concurrent_merge add column x UInt32 default sleep(0.1) on node1 Executing query alter table concurrent_merge add column x UInt32 default sleep(0.1) on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-error_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-error_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-error_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-error_log.xml ] Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_repl' FORMAT Values on replica1 Executing query SELECT database, table, is_permanently, uuid, metadata_path FROM system.detached_tables WHERE database='test_db_repl' FORMAT Values on replica1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-latency_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = Null\n \n \n \n " > /etc/clickhouse-server/config.d/zzz-override-latency_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-latency_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " ENGINE = Null " > /etc/clickhouse-server/config.d/zzz-override-latency_log.xml ] Executing query system start merges on node1 Executing query system start merges on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Executing query DROP DATABASE test_db_repl on replica1 Executing query DROP DATABASE test_db_repl on replica1 Executing query system start merges on node2 Executing query system start merges on node2 Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 2358 ? 00:00:01 clickhouse Stdout: 2358 ? 00:00:01 clickhouse run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] [gw4] PASSED test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query optimize table concurrent_merge final on node1 Executing query optimize table concurrent_merge final on node1 Stdout:2358 Stdout:2358 Request URL: 'http://127.0.0.1:30102/devstoreaccount1/?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '12856c5e-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '12856c5e-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30102 Starting new HTTP connection (1): 127.0.0.1:30102 http://127.0.0.1:30102 "GET /devstoreaccount1/?restype=account&comp=properties HTTP/1.1" 200 0 http://127.0.0.1:30102 "GET /devstoreaccount1/?restype=account&comp=properties HTTP/1.1" 200 0 Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '12856c5e-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': 'eb9fe602-759b-42e1-8aa8-ce62912a1b54' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:12 GMT' 'x-ms-sku-name': 'REDACTED' 'x-ms-account-kind': 'REDACTED' 'x-ms-is-hns-enabled': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '12856c5e-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': 'eb9fe602-759b-42e1-8aa8-ce62912a1b54' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:12 GMT' 'x-ms-sku-name': 'REDACTED' 'x-ms-account-kind': 'REDACTED' 'x-ms-is-hns-enabled': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' {'client_request_id': '12856c5e-0f79-11f0-a196-0242ac110002', 'request_id': 'eb9fe602-759b-42e1-8aa8-ce62912a1b54', 'version': '2025-05-05', 'date': datetime.datetime(2025, 4, 2, 4, 15, 12, tzinfo=datetime.timezone.utc), 'sku_name': 'Standard_RAGRS', 'account_kind': 'StorageV2', 'is_hns_enabled': False} {'client_request_id': '12856c5e-0f79-11f0-a196-0242ac110002', 'request_id': 'eb9fe602-759b-42e1-8aa8-ce62912a1b54', 'version': '2025-05-05', 'date': datetime.datetime(2025, 4, 2, 4, 15, 12, tzinfo=datetime.timezone.utc), 'sku_name': 'Standard_RAGRS', 'account_kind': 'StorageV2', 'is_hns_enabled': False} Request URL: 'http://127.0.0.1:30102/devstoreaccount1/?comp=REDACTED&prefix=REDACTED&include=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '128b4188-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/?comp=REDACTED&prefix=REDACTED&include=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '128b4188-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30102 "GET /devstoreaccount1/?comp=list&prefix=azurite-container&include= HTTP/1.1" 200 None http://127.0.0.1:30102 "GET /devstoreaccount1/?comp=list&prefix=azurite-container&include= HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '128b4188-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': '7982475b-1ff0-4d6a-b0b8-dbda4e71da36' 'x-ms-version': 'REDACTED' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 04:15:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '128b4188-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': '7982475b-1ff0-4d6a-b0b8-dbda4e71da36' 'x-ms-version': 'REDACTED' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 04:15:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Request URL: 'http://127.0.0.1:30102/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '128dfa68-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '128dfa68-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30102 "GET /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 404 None http://127.0.0.1:30102 "GET /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 404 None Response status: 404 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-error-code': 'ContainerNotFound' 'x-ms-request-id': '8fa5a1cb-275b-446d-9af7-6301e68a1ddc' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 04:15:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Response status: 404 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-error-code': 'ContainerNotFound' 'x-ms-request-id': '8fa5a1cb-275b-446d-9af7-6301e68a1ddc' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 04:15:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' azurite container 'azurite-container' doesn't exist, creating it azurite container 'azurite-container' doesn't exist, creating it Request URL: 'http://127.0.0.1:30102/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'PUT' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '128f8996-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'PUT' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '128f8996-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30102 "PUT /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 201 0 http://127.0.0.1:30102 "PUT /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 201 0 Response status: 201 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'etag': '"0x250AC6E85597EE0"' 'last-modified': 'Wed, 02 Apr 2025 04:15:12 GMT' 'x-ms-client-request-id': '128f8996-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': 'ea0dcd7e-c024-40e9-b440-4b24b6c6024f' 'x-ms-version': 'REDACTED' 'Date': 'Wed, 02 Apr 2025 04:15:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' Response status: 201 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'etag': '"0x250AC6E85597EE0"' 'last-modified': 'Wed, 02 Apr 2025 04:15:12 GMT' 'x-ms-client-request-id': '128f8996-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': 'ea0dcd7e-c024-40e9-b440-4b24b6c6024f' 'x-ms-version': 'REDACTED' 'Date': 'Wed, 02 Apr 2025 04:15:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml up -d --no-recreate] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2358 Stdout:2358 Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/.env --project-name roottestsystemdetachedtables-gw4 --file /ClickHouse/tests/integration/test_system_detached_tables/_instances-0-gw4/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Removing Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Removing Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Removed Stderr: Container roottestsystemdetachedtables-gw4-replica1-1 Removed Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Stopping Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Removing Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Removing Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Removing Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Removing Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Stopped Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Removing Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Removing Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Removed Stderr: Container roottestsystemdetachedtables-gw4-zoo2-1 Removed Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Removed Stderr: Container roottestsystemdetachedtables-gw4-zoo1-1 Removed Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Removed Stderr: Container roottestsystemdetachedtables-gw4-zoo3-1 Removed Stderr: Network roottestsystemdetachedtables-gw4_default Removing Stderr: Network roottestsystemdetachedtables-gw4_default Removing Stderr: Network roottestsystemdetachedtables-gw4_default Removed Stderr: Network roottestsystemdetachedtables-gw4_default Removed Cleanup called Cleanup called Docker networks for project roottestsystemdetachedtables-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemdetachedtables-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemdetachedtables-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemdetachedtables-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemdetachedtables-gw4 are DRIVER VOLUME NAME Docker volumes for project roottestsystemdetachedtables-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemdetachedtables-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemdetachedtables-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemdetachedtables-gw4 No running containers for project: roottestsystemdetachedtables-gw4 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Running Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Running Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Creating Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Creating run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Creating Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Creating Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Creating Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Creating Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Created Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Created Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Created Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Created Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Created Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Created Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Starting Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Starting Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Starting Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Starting Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Starting Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Starting Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Started Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Started Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Started Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Started Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Started Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node_0 get_instance_ip instance_name=node_0 http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_0-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_0 get_instance_ip instance_name=node_0 http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_0-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_0, ip: 172.16.1.4... Waiting for ClickHouse start in node_0, ip: 172.16.1.4... http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c7fac1240e87790512effb369d8ab93e6e39b95fd8379ba87790be684b75a1ae/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c7fac1240e87790512effb369d8ab93e6e39b95fd8379ba87790be684b75a1ae/json HTTP/1.1" 200 None Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:2358 Stdout:2358 Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 http://localhost:None "GET /v1.46/containers/c7fac1240e87790512effb369d8ab93e6e39b95fd8379ba87790be684b75a1ae/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c7fac1240e87790512effb369d8ab93e6e39b95fd8379ba87790be684b75a1ae/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c7fac1240e87790512effb369d8ab93e6e39b95fd8379ba87790be684b75a1ae/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c7fac1240e87790512effb369d8ab93e6e39b95fd8379ba87790be684b75a1ae/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c7fac1240e87790512effb369d8ab93e6e39b95fd8379ba87790be684b75a1ae/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c7fac1240e87790512effb369d8ab93e6e39b95fd8379ba87790be684b75a1ae/json HTTP/1.1" 200 None ClickHouse node_0 started ClickHouse node_0 started get_instance_ip instance_name=node_1 get_instance_ip instance_name=node_1 http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_1 get_instance_ip instance_name=node_1 http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_1, ip: 172.16.1.3... Waiting for ClickHouse start in node_1, ip: 172.16.1.3... http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7e024bbc0840ac826b72ca2d906fc41417bcdab6ccca1e983b800d43221c996e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7e024bbc0840ac826b72ca2d906fc41417bcdab6ccca1e983b800d43221c996e/json HTTP/1.1" 200 None ClickHouse node_1 started ClickHouse node_1 started get_instance_ip instance_name=node_2 get_instance_ip instance_name=node_2 http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_2 get_instance_ip instance_name=node_2 http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_2, ip: 172.16.1.5... Waiting for ClickHouse start in node_2, ip: 172.16.1.5... http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorageazureblobstoragecluster-gw2-node_2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cd06acd545a0f5b100773bc4b4d374995f99921fc79b8d44f268b7477e718ff7/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cd06acd545a0f5b100773bc4b4d374995f99921fc79b8d44f268b7477e718ff7/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cd06acd545a0f5b100773bc4b4d374995f99921fc79b8d44f268b7477e718ff7/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cd06acd545a0f5b100773bc4b4d374995f99921fc79b8d44f268b7477e718ff7/json HTTP/1.1" 200 None ClickHouse node_2 started ClickHouse node_2 started Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_with_named_collection.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'auto', 'auto', 'a UInt64') VALUES (1), (2) on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_with_named_collection.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'auto', 'auto', 'a UInt64') VALUES (1), (2) on node_0 Executing query SELECT * from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_with_named_collection.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node_0 Executing query SELECT * from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_with_named_collection.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node_0 Executing query SELECT * from azureBlobStorageCluster('simple_cluster', azure_conf2, storage_account_url = 'http://azurite1:30102/devstoreaccount1', container='cont', blob_path='test_cluster_with_named_collection.csv') on node_0 Executing query SELECT * from azureBlobStorageCluster('simple_cluster', azure_conf2, storage_account_url = 'http://azurite1:30102/devstoreaccount1', container='cont', blob_path='test_cluster_with_named_collection.csv') on node_0 [gw2] PASSED test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection test_storage_azure_blob_storage/test_cluster.py::test_count Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_count.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV', 'auto', 'key UInt64') VALUES (1), (2) on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_count.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV', 'auto', 'key UInt64') VALUES (1), (2) on node_0 Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_cluster_count.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '13e937b0-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_cluster_count.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '13e937b0-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30102 Starting new HTTP connection (1): 127.0.0.1:30102 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_cluster_count.csv HTTP/1.1" 206 4 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_cluster_count.csv HTTP/1.1" 206 4 Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:14 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '4' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x235A29111034F00"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '13e937b0-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': 'e712fd3d-6017-4268-ae3e-ca71d656e1b3' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:14 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:14 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '4' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x235A29111034F00"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '13e937b0-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': 'e712fd3d-6017-4268-ae3e-ca71d656e1b3' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:14 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Executing query SELECT count(*) from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_count.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV','auto', 'key UInt64') on node_0 Executing query SELECT count(*) from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_count.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV','auto', 'key UInt64') on node_0 Executing query SELECT count(*) from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_count.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV','auto', 'key UInt64') on node_0 Executing query SELECT count(*) from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_count.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV','auto', 'key UInt64') on node_0 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 [gw2] PASSED test_storage_azure_blob_storage/test_cluster.py::test_count test_storage_azure_blob_storage/test_cluster.py::test_format_detection Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection0', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'JSONEachRow', 'auto', 'x UInt32, y String') select number as x, 'str_' || toString(number) from numbers(10) SETTINGS azure_truncate_on_insert=1 on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection0', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'JSONEachRow', 'auto', 'x UInt32, y String') select number as x, 'str_' || toString(number) from numbers(10) SETTINGS azure_truncate_on_insert=1 on node_0 http://localhost:None "POST /v1.46/exec/dad834593c2556565a3a6e7bb52d60df9c6058d42cffa67b30486e32dda461c4/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/dad834593c2556565a3a6e7bb52d60df9c6058d42cffa67b30486e32dda461c4/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/dad834593c2556565a3a6e7bb52d60df9c6058d42cffa67b30486e32dda461c4/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/dad834593c2556565a3a6e7bb52d60df9c6058d42cffa67b30486e32dda461c4/json HTTP/1.1" 200 586 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection1', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'JSONEachRow', 'auto', 'x UInt32, y String') select number as x, 'str_' || toString(number) from numbers(10, 10) SETTINGS azure_truncate_on_insert=1 on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection1', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'JSONEachRow', 'auto', 'x UInt32, y String') select number as x, 'str_' || toString(number) from numbers(10, 10) SETTINGS azure_truncate_on_insert=1 on node_0 Executing query desc azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'JSONEachRow', 'auto', 'auto') on node_0 Executing query desc azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'JSONEachRow', 'auto', 'auto') on node_0 Executing query desc azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node_0 Executing query desc azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node_0 Executing query select * from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'JSONEachRow', 'auto', 'x UInt32, y String') order by x on node_0 Executing query select * from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'JSONEachRow', 'auto', 'x UInt32, y String') order by x on node_0 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') order by x on node_0 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') order by x on node_0 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', auto) order by x on node_0 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', auto) order by x on node_0 Executing query select count() from system.parts where table = 'concurrent_merge' and active on node1 Executing query select count() from system.parts where table = 'concurrent_merge' and active on node1 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', auto, auto) order by x on node_0 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', auto, auto) order by x on node_0 Executing query select count() from system.parts where table = 'concurrent_merge' and active on node2 Executing query select count() from system.parts where table = 'concurrent_merge' and active on node2 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:3185 Stdout:3185 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'x UInt32, y String') order by x on node_0 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'x UInt32, y String') order by x on node_0 Stdout:3185 Stdout:3185 Executing query select 20 on node_default Executing query select 20 on node_default Executing query SYSTEM FLUSH LOGS on node_default Executing query SYSTEM FLUSH LOGS on node_default Executing query select count() from system.parts where table = 'concurrent_merge' and active on node2 Executing query select count() from system.parts where table = 'concurrent_merge' and active on node2 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', auto, auto, 'x UInt32, y String') order by x on node_0 Executing query select * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_format_detection*', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', auto, auto, 'x UInt32, y String') order by x on node_0 Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SHOW CREATE TABLE system.query_log on node_default [gw2] PASSED test_storage_azure_blob_storage/test_cluster.py::test_format_detection test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_tf_{_partition_id}.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV', 'auto', 'column1 UInt32, column2 UInt32, column3 UInt32') PARTITION BY column3 VALUES (1, 2, 3), (3, 2, 1), (78, 43, 45) on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_tf_{_partition_id}.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV', 'auto', 'column1 UInt32, column2 UInt32, column3 UInt32') PARTITION BY column3 VALUES (1, 2, 3), (3, 2, 1), (78, 43, 45) on node_0 Executing query select count() from system.parts where table = 'concurrent_merge' and active on node2 Executing query select count() from system.parts where table = 'concurrent_merge' and active on node2 Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_tf_3.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '14ef0c0c-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_tf_3.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '14ef0c0c-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30102 Starting new HTTP connection (1): 127.0.0.1:30102 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_tf_3.csv HTTP/1.1" 206 6 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_tf_3.csv HTTP/1.1" 206 6 Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '6' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x1FE67E24B3006B0"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '14ef0c0c-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': '663937e6-8ae0-4ce8-8bf8-add387da31c5' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '6' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x1FE67E24B3006B0"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '14ef0c0c-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': '663937e6-8ae0-4ce8-8bf8-add387da31c5' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_tf_1.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '14f039e2-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_tf_1.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '14f039e2-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30102 Starting new HTTP connection (1): 127.0.0.1:30102 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_tf_1.csv HTTP/1.1" 206 6 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_tf_1.csv HTTP/1.1" 206 6 Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '6' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x1C663CF15BD7AB0"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '14f039e2-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': '2355ce0b-0f54-447b-9cc8-9df89edc69e6' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '6' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x1C663CF15BD7AB0"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '14f039e2-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': '2355ce0b-0f54-447b-9cc8-9df89edc69e6' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_tf_45.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '14f121fe-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_tf_45.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '14f121fe-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30102 Starting new HTTP connection (1): 127.0.0.1:30102 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_tf_45.csv HTTP/1.1" 206 9 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_tf_45.csv HTTP/1.1" 206 9 Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '9' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x1BF7D44363C74B0"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '14f121fe-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': 'ce71f172-7790-4072-91cc-6fdfe7f56541' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '9' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x1BF7D44363C74B0"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '14f121fe-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': 'ce71f172-7790-4072-91cc-6fdfe7f56541' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Executing query SELECT count(*) from azureBlobStorageCluster('simple_cluster', azure_conf2, storage_account_url = 'http://azurite1:30102/devstoreaccount1', container='cont', blob_path='test_tf_*.csv', format='CSV', compression='auto', structure='column1 UInt32, column2 UInt32, column3 UInt32') on node_0 Executing query SELECT count(*) from azureBlobStorageCluster('simple_cluster', azure_conf2, storage_account_url = 'http://azurite1:30102/devstoreaccount1', container='cont', blob_path='test_tf_*.csv', format='CSV', compression='auto', structure='column1 UInt32, column2 UInt32, column3 UInt32') on node_0 Executing query select sum(id) from concurrent_merge on node1 Executing query select sum(id) from concurrent_merge on node1 Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query select sum(id) from concurrent_merge on node2 Executing query select sum(id) from concurrent_merge on node2 Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query SHOW CREATE TABLE system.query_metric_log on node_default [gw2] PASSED test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster test_storage_azure_blob_storage/test_cluster.py::test_select_all Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_select_all.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV', 'auto', 'key UInt64, data String') VALUES (1, 'a'), (2, 'b') on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_select_all.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV', 'auto', 'key UInt64, data String') VALUES (1, 'a'), (2, 'b') on node_0 Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_cluster_select_all.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '151d21d2-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Request URL: 'http://127.0.0.1:30102/devstoreaccount1/cont/test_cluster_select_all.csv' Request method: 'GET' Request headers: 'x-ms-range': 'REDACTED' 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '151d21d2-0f79-11f0-a196-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30102 Starting new HTTP connection (1): 127.0.0.1:30102 [gw0] http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_cluster_select_all.csv HTTP/1.1" 206 12 http://127.0.0.1:30102 "GET /devstoreaccount1/cont/test_cluster_select_all.csv HTTP/1.1" 206 12 PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '12' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x239703A08328E80"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '151d21d2-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': '39c7f175-e386-4cfb-9dae-2b3b8bda7f7b' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' Response status: 206 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'last-modified': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-creation-time': 'REDACTED' 'content-length': '12' 'content-type': 'application/octet-stream' 'content-range': 'REDACTED' 'etag': '"0x239703A08328E80"' 'x-ms-blob-type': 'REDACTED' 'x-ms-lease-state': 'REDACTED' 'x-ms-lease-status': 'REDACTED' 'x-ms-client-request-id': '151d21d2-0f79-11f0-a196-0242ac110002' 'x-ms-request-id': '39c7f175-e386-4cfb-9dae-2b3b8bda7f7b' 'x-ms-version': 'REDACTED' 'accept-ranges': 'REDACTED' 'date': 'Wed, 02 Apr 2025 04:15:16 GMT' 'x-ms-server-encrypted': 'REDACTED' 'x-ms-blob-content-md5': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter Executing query DROP TABLE IF EXISTS drop_detached_test SYNC on node1 Executing query DROP TABLE IF EXISTS drop_detached_test SYNC on node1 Executing query SELECT * from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_select_all.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV','auto') on node_0 Executing query SELECT * from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_select_all.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV','auto') on node_0 Executing query SHOW CREATE TABLE system.query_thread_log on node_default Executing query SHOW CREATE TABLE system.query_thread_log on node_default Stderr: Container roottests3withhttps-gw1-resolver-1 Stopping Stderr: Container roottests3withhttps-gw1-resolver-1 Stopping Stderr: Container roottests3withhttps-gw1-node-1 Stopping Stderr: Container roottests3withhttps-gw1-node-1 Stopping Stderr: Container roottests3withhttps-gw1-node-1 Stopped Stderr: Container roottests3withhttps-gw1-node-1 Stopped Stderr: Container roottests3withhttps-gw1-minio1-1 Stopping Stderr: Container roottests3withhttps-gw1-minio1-1 Stopping Stderr: Container roottests3withhttps-gw1-minio1-1 Stopped Stderr: Container roottests3withhttps-gw1-minio1-1 Stopped Stderr: Container roottests3withhttps-gw1-resolver-1 Stopped Stderr: Container roottests3withhttps-gw1-resolver-1 Stopped Stderr: Container roottests3withhttps-gw1-proxy1-1 Stopping Stderr: Container roottests3withhttps-gw1-proxy1-1 Stopping Stderr: Container roottests3withhttps-gw1-proxy2-1 Stopping Stderr: Container roottests3withhttps-gw1-proxy2-1 Stopping Stderr: Container roottests3withhttps-gw1-proxy1-1 Stopped Stderr: Container roottests3withhttps-gw1-proxy1-1 Stopped Stderr: Container roottests3withhttps-gw1-proxy2-1 Stopped Stderr: Container roottests3withhttps-gw1-proxy2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/.env --project-name roottests3withhttps-gw1 --file /ClickHouse/tests/integration/test_s3_with_https/_instances-0-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Executing query DROP TABLE IF EXISTS drop_detached_test SYNC on node2 Executing query DROP TABLE IF EXISTS drop_detached_test SYNC on node2 Executing query SELECT * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_select_all.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV','auto') on node_0 Executing query SELECT * from azureBlobStorageCluster('simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_cluster_select_all.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'CSV','auto') on node_0 Executing query SHOW CREATE TABLE system.query_thread_log on node_default Executing query SHOW CREATE TABLE system.query_thread_log on node_default Executing query CREATE TABLE drop_detached_test ON CLUSTER test_cluster (d UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/drop_detached_test', '{replica}') ORDER BY d PARTITION BY d SETTINGS storage_policy='s3' on node1 Executing query CREATE TABLE drop_detached_test ON CLUSTER test_cluster (d UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/drop_detached_test', '{replica}') ORDER BY d PARTITION BY d SETTINGS storage_policy='s3' on node1 Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default [gw2] PASSED test_storage_azure_blob_storage/test_cluster.py::test_select_all test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_skip_unavailable.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'auto', 'auto', 'a UInt64') VALUES (1), (2) on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_skip_unavailable.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'auto', 'auto', 'a UInt64') VALUES (1), (2) on node_0 Executing query INSERT INTO drop_detached_test VALUES (0) on node1 Executing query INSERT INTO drop_detached_test VALUES (0) on node1 Executing query SHOW CREATE TABLE system.part_log on node_default Executing query SHOW CREATE TABLE system.part_log on node_default Executing query SELECT count(*) from azureBlobStorageCluster('cluster_non_existent_port','http://azurite1:30102/devstoreaccount1', 'cont', 'test_skip_unavailable.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') SETTINGS skip_unavailable_shards = 1 on node_0 Executing query SELECT count(*) from azureBlobStorageCluster('cluster_non_existent_port','http://azurite1:30102/devstoreaccount1', 'cont', 'test_skip_unavailable.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') SETTINGS skip_unavailable_shards = 1 on node_0 Executing query ALTER TABLE drop_detached_test FREEZE WITH NAME 'detach_backup1' on node1 Executing query ALTER TABLE drop_detached_test FREEZE WITH NAME 'detach_backup1' on node1 Executing query SHOW CREATE TABLE system.part_log on node_default Executing query SHOW CREATE TABLE system.part_log on node_default Executing query INSERT INTO drop_detached_test VALUES (1) on node1 Executing query INSERT INTO drop_detached_test VALUES (1) on node1 [gw2] PASSED test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default test_storage_azure_blob_storage/test_cluster.py::test_union_all Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') VALUES (1, 'a'), (2, 'b'), (3, 'c'), (4, 'd') on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') VALUES (1, 'a'), (2, 'b'), (3, 'c'), (4, 'd') on node_0 Executing query ALTER TABLE drop_detached_test FREEZE WITH NAME 'detach_backup2' on node1 Executing query ALTER TABLE drop_detached_test FREEZE WITH NAME 'detach_backup2' on node1 Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SELECT * FROM ( SELECT * from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') UNION ALL SELECT * from azureBlobStorage( 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') ) ORDER BY (a) on node_0 Executing query SELECT * FROM ( SELECT * from azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') UNION ALL SELECT * from azureBlobStorage( 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') ) ORDER BY (a) on node_0 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SELECT * FROM ( SELECT * from azureBlobStorageCluster( 'simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') UNION ALL SELECT * from azureBlobStorageCluster( 'simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') ) ORDER BY (a) on node_0 Executing query SELECT * FROM ( SELECT * from azureBlobStorageCluster( 'simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') UNION ALL SELECT * from azureBlobStorageCluster( 'simple_cluster', 'http://azurite1:30102/devstoreaccount1', 'cont', 'test_parquet_union_all', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'Parquet', 'auto', 'a Int32, b String') ) ORDER BY (a) on node_0 run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/detach_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/detach_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Stderr: Container roottests3withhttps-gw1-node-1 Stopping Stderr: Container roottests3withhttps-gw1-node-1 Stopping Stderr: Container roottests3withhttps-gw1-resolver-1 Stopping Stderr: Container roottests3withhttps-gw1-resolver-1 Stopping Stderr: Container roottests3withhttps-gw1-node-1 Stopped Stderr: Container roottests3withhttps-gw1-node-1 Stopped Stderr: Container roottests3withhttps-gw1-node-1 Removing Stderr: Container roottests3withhttps-gw1-node-1 Removing Stderr: Container roottests3withhttps-gw1-resolver-1 Stopped Stderr: Container roottests3withhttps-gw1-resolver-1 Stopped Stderr: Container roottests3withhttps-gw1-resolver-1 Removing Stderr: Container roottests3withhttps-gw1-resolver-1 Removing Stderr: Container roottests3withhttps-gw1-resolver-1 Removed Stderr: Container roottests3withhttps-gw1-resolver-1 Removed Stderr: Container roottests3withhttps-gw1-node-1 Removed Stderr: Container roottests3withhttps-gw1-node-1 Removed Stderr: Container roottests3withhttps-gw1-minio1-1 Stopping Stderr: Container roottests3withhttps-gw1-minio1-1 Stopping Stderr: Container roottests3withhttps-gw1-minio1-1 Stopped Stderr: Container roottests3withhttps-gw1-minio1-1 Stopped Stderr: Container roottests3withhttps-gw1-minio1-1 Removing Stderr: Container roottests3withhttps-gw1-minio1-1 Removing Stderr: Container roottests3withhttps-gw1-minio1-1 Removed Stderr: Container roottests3withhttps-gw1-minio1-1 Removed Stderr: Container roottests3withhttps-gw1-proxy1-1 Stopping Stderr: Container roottests3withhttps-gw1-proxy1-1 Stopping Stderr: Container roottests3withhttps-gw1-proxy2-1 Stopping Stderr: Container roottests3withhttps-gw1-proxy2-1 Stopping Stderr: Container roottests3withhttps-gw1-proxy1-1 Stopped Stderr: Container roottests3withhttps-gw1-proxy1-1 Stopped Stderr: Container roottests3withhttps-gw1-proxy1-1 Removing Stderr: Container roottests3withhttps-gw1-proxy1-1 Removing Stderr: Container roottests3withhttps-gw1-proxy2-1 Stopped Stderr: Container roottests3withhttps-gw1-proxy2-1 Stopped Stderr: Container roottests3withhttps-gw1-proxy2-1 Removing Stderr: Container roottests3withhttps-gw1-proxy2-1 Removing Stderr: Container roottests3withhttps-gw1-proxy1-1 Removed Stderr: Container roottests3withhttps-gw1-proxy1-1 Removed Stderr: Container roottests3withhttps-gw1-proxy2-1 Removed Stderr: Container roottests3withhttps-gw1-proxy2-1 Removed Stderr: Volume roottests3withhttps-gw1_data1-1 Removing Stderr: Volume roottests3withhttps-gw1_data1-1 Removing Stderr: Network roottests3withhttps-gw1_default Removing Stderr: Network roottests3withhttps-gw1_default Removing Stderr: Volume roottests3withhttps-gw1_data1-1 Removed Stderr: Volume roottests3withhttps-gw1_data1-1 Removed Stderr: Network roottests3withhttps-gw1_default Removed Stderr: Network roottests3withhttps-gw1_default Removed Cleanup called Cleanup called Docker networks for project roottests3withhttps-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3withhttps-gw1 are NETWORK ID NAME DRIVER SCOPE [gw2] PASSED test_storage_azure_blob_storage/test_cluster.py::test_union_all test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_unset_skip_unavailable.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'auto', 'auto', 'a UInt64') VALUES (1), (2) on node_0 Executing query INSERT INTO TABLE FUNCTION azureBlobStorage('http://azurite1:30102/devstoreaccount1', 'cont', 'test_unset_skip_unavailable.csv', 'devstoreaccount1', 'Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==', 'auto', 'auto', 'a UInt64') VALUES (1), (2) on node_0 Executing query SHOW CREATE TABLE system.metric_log on node_default Executing query SHOW CREATE TABLE system.metric_log on node_default Docker containers for project roottests3withhttps-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3withhttps-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottests3withhttps-gw1 are DRIVER VOLUME NAME Docker volumes for project roottests3withhttps-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottests3withhttps-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottests3withhttps-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottests3withhttps-gw1 No running containers for project: roottests3withhttps-gw1 Trying to prune unused networks... Trying to prune unused networks... run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/detach_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/detach_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Stdout:11 Stdout:11 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 11 Volumes pruned: 11 test_store_cleanup/test.py::test_store_cleanup Running tests in /ClickHouse/tests/integration/test_store_cleanup/test.py Running tests in /ClickHouse/tests/integration/test_store_cleanup/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project rootteststorecleanup-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorecleanup-gw1 are NETWORK ID NAME DRIVER SCOPE Executing query ALTER TABLE drop_detached_test UNFREEZE WITH NAME 'detach_backup2' on node1 Executing query ALTER TABLE drop_detached_test UNFREEZE WITH NAME 'detach_backup2' on node1 Docker containers for project rootteststorecleanup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorecleanup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query SELECT count(*) from azureBlobStorageCluster('cluster_non_existent_port','http://azurite1:30102/devstoreaccount1', 'cont', 'test_unset_skip_unavailable.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node_0 Executing query SELECT count(*) from azureBlobStorageCluster('cluster_non_existent_port','http://azurite1:30102/devstoreaccount1', 'cont', 'test_unset_skip_unavailable.csv', 'devstoreaccount1','Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==') on node_0 Executing query SHOW CREATE TABLE system.metric_log on node_default Executing query SHOW CREATE TABLE system.metric_log on node_default Docker volumes for project rootteststorecleanup-gw1 are DRIVER VOLUME NAME Docker volumes for project rootteststorecleanup-gw1 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project rootteststorecleanup-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorecleanup-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststorecleanup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorecleanup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query ALTER TABLE drop_detached_test UNFREEZE WITH NAME 'detach_backup1' on node1 Executing query ALTER TABLE drop_detached_test UNFREEZE WITH NAME 'detach_backup1' on node1 Docker volumes for project rootteststorecleanup-gw1 are DRIVER VOLUME NAME Docker volumes for project rootteststorecleanup-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorecleanup-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorecleanup-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorecleanup-gw1 No running containers for project: rootteststorecleanup-gw1 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml stop --timeout 20] [gw2] PASSED test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Executing query ALTER TABLE drop_detached_test DETACH PARTITION '0' on node1 Executing query ALTER TABLE drop_detached_test DETACH PARTITION '0' on node1 Stdout:11 Stdout:11 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query SHOW CREATE TABLE system.error_log on node_default Executing query SHOW CREATE TABLE system.error_log on node_default Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 11 Volumes pruned: 11 Setup directory for instance: node1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_store_cleanup/configs/store_cleanup.xml'] to /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_store_cleanup/configs/store_cleanup.xml'] to /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/database Setup database dir /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/database Setup logs dir /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/logs Setup logs dir /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml pull] Executing query ALTER TABLE drop_detached_test DETACH PARTITION '1' on node1 Executing query ALTER TABLE drop_detached_test DETACH PARTITION '1' on node1 Executing query SHOW CREATE TABLE system.error_log on node_default Executing query SHOW CREATE TABLE system.error_log on node_default Stderr: Container roottestthrottling-gw7-node-1 Stopping Stderr: Container roottestthrottling-gw7-node-1 Stopping Stderr: Container roottestthrottling-gw7-resolver-1 Stopping Stderr: Container roottestthrottling-gw7-resolver-1 Stopping Stderr: Container roottestthrottling-gw7-node-1 Stopped Stderr: Container roottestthrottling-gw7-node-1 Stopped Stderr: Container roottestthrottling-gw7-minio1-1 Stopping Stderr: Container roottestthrottling-gw7-minio1-1 Stopping Stderr: Container roottestthrottling-gw7-minio1-1 Stopped Stderr: Container roottestthrottling-gw7-minio1-1 Stopped Stderr: Container roottestthrottling-gw7-resolver-1 Stopped Stderr: Container roottestthrottling-gw7-resolver-1 Stopped Stderr: Container roottestthrottling-gw7-proxy2-1 Stopping Stderr: Container roottestthrottling-gw7-proxy2-1 Stopping Stderr: Container roottestthrottling-gw7-proxy1-1 Stopping Stderr: Container roottestthrottling-gw7-proxy1-1 Stopping Stderr: Container roottestthrottling-gw7-proxy2-1 Stopped Stderr: Container roottestthrottling-gw7-proxy2-1 Stopped Stderr: Container roottestthrottling-gw7-proxy1-1 Stopped Stderr: Container roottestthrottling-gw7-proxy1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Command:[docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/.env --project-name roottestthrottling-gw7 --file /ClickHouse/tests/integration/test_throttling/_instances-0-gw7/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-query_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-query_log.xml ] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml ] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml ] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-part_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-part_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-part_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-part_log.xml ] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-trace_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-trace_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-trace_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-trace_log.xml ] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-metric_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-metric_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-metric_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-metric_log.xml ] Stderr: Container roottestthrottling-gw7-node-1 Stopping Stderr: Container roottestthrottling-gw7-node-1 Stopping Stderr: Container roottestthrottling-gw7-resolver-1 Stopping Stderr: Container roottestthrottling-gw7-resolver-1 Stopping Stderr: Container roottestthrottling-gw7-node-1 Stopped Stderr: Container roottestthrottling-gw7-node-1 Stopped Stderr: Container roottestthrottling-gw7-node-1 Removing Stderr: Container roottestthrottling-gw7-node-1 Removing Stderr: Container roottestthrottling-gw7-resolver-1 Stopped Stderr: Container roottestthrottling-gw7-resolver-1 Stopped Stderr: Container roottestthrottling-gw7-resolver-1 Removing Stderr: Container roottestthrottling-gw7-resolver-1 Removing Stderr: Container roottestthrottling-gw7-node-1 Removed Stderr: Container roottestthrottling-gw7-node-1 Removed Stderr: Container roottestthrottling-gw7-minio1-1 Stopping Stderr: Container roottestthrottling-gw7-minio1-1 Stopping Stderr: Container roottestthrottling-gw7-resolver-1 Removed Stderr: Container roottestthrottling-gw7-resolver-1 Removed Stderr: Container roottestthrottling-gw7-minio1-1 Stopped Stderr: Container roottestthrottling-gw7-minio1-1 Stopped Stderr: Container roottestthrottling-gw7-minio1-1 Removing Stderr: Container roottestthrottling-gw7-minio1-1 Removing Stderr: Container roottestthrottling-gw7-minio1-1 Removed Stderr: Container roottestthrottling-gw7-minio1-1 Removed Stderr: Container roottestthrottling-gw7-proxy1-1 Stopping Stderr: Container roottestthrottling-gw7-proxy1-1 Stopping Stderr: Container roottestthrottling-gw7-proxy2-1 Stopping Stderr: Container roottestthrottling-gw7-proxy2-1 Stopping Stderr: Container roottestthrottling-gw7-proxy1-1 Stopped Stderr: Container roottestthrottling-gw7-proxy1-1 Stopped Stderr: Container roottestthrottling-gw7-proxy1-1 Removing Stderr: Container roottestthrottling-gw7-proxy1-1 Removing Stderr: Container roottestthrottling-gw7-proxy2-1 Stopped Stderr: Container roottestthrottling-gw7-proxy2-1 Stopped Stderr: Container roottestthrottling-gw7-proxy2-1 Removing Stderr: Container roottestthrottling-gw7-proxy2-1 Removing Stderr: Container roottestthrottling-gw7-proxy1-1 Removed Stderr: Container roottestthrottling-gw7-proxy1-1 Removed Stderr: Container roottestthrottling-gw7-proxy2-1 Removed Stderr: Container roottestthrottling-gw7-proxy2-1 Removed Stderr: Volume roottestthrottling-gw7_data1-1 Removing Stderr: Volume roottestthrottling-gw7_data1-1 Removing Stderr: Network roottestthrottling-gw7_default Removing Stderr: Network roottestthrottling-gw7_default Removing Stderr: Volume roottestthrottling-gw7_data1-1 Removed Stderr: Volume roottestthrottling-gw7_data1-1 Removed Stderr: Network roottestthrottling-gw7_default Removed Stderr: Network roottestthrottling-gw7_default Removed Cleanup called Cleanup called run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-error_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-error_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-error_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-error_log.xml ] Docker networks for project roottestthrottling-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestthrottling-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestthrottling-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestthrottling-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-latency_log.xml\n '] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n system_tables\n \n \n " > /etc/clickhouse-server/config.d/zzz-override-latency_log.xml\n '] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-latency_log.xml ] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c echo " system_tables " > /etc/clickhouse-server/config.d/zzz-override-latency_log.xml ] Docker volumes for project roottestthrottling-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestthrottling-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestthrottling-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestthrottling-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestthrottling-gw7 No running containers for project: roottestthrottling-gw7 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:9 Stdout:9 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 3185 ? 00:00:01 clickhouse Stdout: 3185 ? 00:00:01 clickhouse run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 9 Volumes pruned: 9 test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue Running tests in /ClickHouse/tests/integration/test_system_ddl_worker_queue/test.py Running tests in /ClickHouse/tests/integration/test_system_ddl_worker_queue/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottestsystemddlworkerqueue-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemddlworkerqueue-gw7 are NETWORK ID NAME DRIVER SCOPE run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Docker containers for project roottestsystemddlworkerqueue-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemddlworkerqueue-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemddlworkerqueue-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemddlworkerqueue-gw7 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottestsystemddlworkerqueue-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemddlworkerqueue-gw7 are NETWORK ID NAME DRIVER SCOPE Stdout:3185 Stdout:3185 Docker containers for project roottestsystemddlworkerqueue-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemddlworkerqueue-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemddlworkerqueue-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemddlworkerqueue-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemddlworkerqueue-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemddlworkerqueue-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemddlworkerqueue-gw7 No running containers for project: roottestsystemddlworkerqueue-gw7 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:9 Stdout:9 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 9 Volumes pruned: 9 Setup directory for instance: node1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/database Setup database dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/database Setup logs dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/logs Setup logs dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/database Setup database dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/database Setup logs dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/logs Setup logs dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node3 Setup directory for instance: node3 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/database Setup database dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/database Setup logs dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/logs Setup logs dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node4 Setup directory for instance: node4 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/database Setup database dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/database Setup logs dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/logs Setup logs dir /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml pull] Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:3185 Stdout:3185 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '1' on node2 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '1' on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node1 Executing query SYSTEM SYNC REPLICA drop_detached_test on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr: node4 Skipped - Image is already being pulled by node3 Stderr: node4 Skipped - Image is already being pulled by node3 Stderr: node1 Skipped - Image is already being pulled by node3 Stderr: node1 Skipped - Image is already being pulled by node3 Stderr: zoo2 Skipped - Image is already being pulled by node3 Stderr: zoo2 Skipped - Image is already being pulled by node3 Stderr: zoo3 Skipped - Image is already being pulled by node3 Stderr: zoo3 Skipped - Image is already being pulled by node3 Stderr: zoo1 Skipped - Image is already being pulled by node3 Stderr: zoo1 Skipped - Image is already being pulled by node3 Stderr: node2 Skipped - Image is already being pulled by node3 Stderr: node2 Skipped - Image is already being pulled by node3 Stderr: node3 Pulling Stderr: node3 Pulling Stderr: node3 Pulled Stderr: node3 Pulled Setup ZooKeeper Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/log', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/config', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/coordination', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/log', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/config', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/coordination', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/log', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/config', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/coordination'] Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/log', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/config', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper1/coordination', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/log', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/config', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper2/coordination', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/log', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/config', '/ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/keeper3/coordination'] Command:[docker compose --project-name roottestsystemddlworkerqueue-gw7 --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Command:[docker compose --project-name roottestsystemddlworkerqueue-gw7 --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stdout:3185 Stdout:3185 Executing query SELECT * FROM system.merges WHERE table = 'mutation_simple' and progress < 1 on node1 Executing query SELECT * FROM system.merges WHERE table = 'mutation_simple' and progress < 1 on node1 Stderr: node1 Pulling Stderr: node1 Pulling Stderr: node1 Pulled Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml up -d --no-recreate] Executing query SELECT * FROM system.merges WHERE table = 'mutation_simple' and progress < 1 on node1 Executing query SELECT * FROM system.merges WHERE table = 'mutation_simple' and progress < 1 on node1 Executing query DROP TABLE test.mutation_simple on node1 Executing query DROP TABLE test.mutation_simple on node1 Stderr: mongo_no_cred Skipped - Image is already being pulled by mongo1 Stderr: mongo_no_cred Skipped - Image is already being pulled by mongo1 Stderr: mongo_secure Skipped - Image is already being pulled by mongo1 Stderr: mongo_secure Skipped - Image is already being pulled by mongo1 Stderr: node Pulling Stderr: node Pulling Stderr: mongo1 Pulling Stderr: mongo1 Pulling Stderr: node Pulled Stderr: node Pulled Stderr: mongo1 Pulled Stderr: mongo1 Pulled Setup Mongo Setup Mongo Command:[docker compose --project-name roottesttablefunctionmongodblegacy-gw3 --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml --verbose up -d] Command:[docker compose --project-name roottesttablefunctionmongodblegacy-gw3 --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml --verbose up -d] [gw6] PASSED test_system_merges/test.py::test_mutation_simple[] test_system_merges/test.py::test_mutation_simple[replicated] Executing query create table test.mutation_simple (a Int64) engine=ReplicatedMergeTree('/clickhouse/test_mutation_simple', '{replica}') order by tuple() on node1 Executing query create table test.mutation_simple (a Int64) engine=ReplicatedMergeTree('/clickhouse/test_mutation_simple', '{replica}') order by tuple() on node1 Executing query create table test.mutation_simple (a Int64) engine=ReplicatedMergeTree('/clickhouse/test_mutation_simple', '{replica}') order by tuple() on node2 Executing query create table test.mutation_simple (a Int64) engine=ReplicatedMergeTree('/clickhouse/test_mutation_simple', '{replica}') order by tuple() on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query INSERT INTO test.mutation_simple VALUES (1), (2), (3) on node1 Executing query INSERT INTO test.mutation_simple VALUES (1), (2), (3) on node1 Stderr:time="2025-04-02T04:15:21Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:15:21Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestsystemddlworkerqueue-gw7_default Creating Stderr: Network roottestsystemddlworkerqueue-gw7_default Creating Stderr: Network roottestsystemddlworkerqueue-gw7_default Created Stderr: Network roottestsystemddlworkerqueue-gw7_default Created Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Started Stderr:time="2025-04-02T04:15:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:22Z" level=debug msg="otel error" error="" Wait ZooKeeper to start Wait ZooKeeper to start get_instance_ip instance_name=zoo1 get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-zoo1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.4, port:2181, use_ssl:False get_kazoo_client: zoo1, ip:172.16.2.4, port:2181, use_ssl:False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query ALTER TABLE test.mutation_simple UPDATE a = 42 WHERE sleep(9) = 0 on node1 Executing query ALTER TABLE test.mutation_simple UPDATE a = 42 WHERE sleep(9) = 0 on node1 Executing query select count() from system.merges where table='mutation_simple' on node2 Executing query select count() from system.merges where table='mutation_simple' on node2 Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr: Network rootteststorecleanup-gw1_default Creating Stderr: Network rootteststorecleanup-gw1_default Creating Stderr: Network rootteststorecleanup-gw1_default Created Stderr: Network rootteststorecleanup-gw1_default Created Stderr: Container rootteststorecleanup-gw1-node1-1 Creating Stderr: Container rootteststorecleanup-gw1-node1-1 Creating Stderr: Container rootteststorecleanup-gw1-node1-1 Created Stderr: Container rootteststorecleanup-gw1-node1-1 Created Stderr: Container rootteststorecleanup-gw1-node1-1 Starting Stderr: Container rootteststorecleanup-gw1-node1-1 Starting Stderr: Container rootteststorecleanup-gw1-node1-1 Started Stderr: Container rootteststorecleanup-gw1-node1-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/rootteststorecleanup-gw1-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorecleanup-gw1-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/rootteststorecleanup-gw1-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorecleanup-gw1-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.7.2... Waiting for ClickHouse start in node1, ip: 172.16.7.2... http://localhost:None "GET /v1.46/containers/rootteststorecleanup-gw1-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststorecleanup-gw1-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None Stdout:3185 Stdout:3185 run container_id:roottests3tablefunctions-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'tc qdisc del dev eth0 root netem'] run container_id:roottests3tablefunctions-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'tc qdisc del dev eth0 root netem'] Command:[docker exec -u root roottests3tablefunctions-gw5-node-1 bash -c tc qdisc del dev eth0 root netem] Command:[docker exec -u root roottests3tablefunctions-gw5-node-1 bash -c tc qdisc del dev eth0 root netem] http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None Executing query select count() from system.merges where table='mutation_simple' on node2 Executing query select count() from system.merges where table='mutation_simple' on node2 Stopping cluster Stopping cluster Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] [gw5] PASSED test_s3_table_functions/test.py::test_s3_table_functions_timeouts Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None Executing query SELECT database, table, num_parts, source_part_names, source_part_paths, result_part_name, result_part_path, partition_id, is_mutation FROM system.merges WHERE table = 'mutation_simple' on node2 Executing query SELECT database, table, num_parts, source_part_names, source_part_paths, result_part_name, result_part_path, partition_id, is_mutation FROM system.merges WHERE table = 'mutation_simple' on node2 Stderr:time="2025-04-02T04:15:21Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:15:21Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Creating Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Creating Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Created Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Started Stderr:time="2025-04-02T04:15:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:22Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:22Z" level=debug msg="otel error" error="" http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/30c284d1a61b97c380212ba8ca8c436ebbf51111b76768386b4d3643571ff806/json HTTP/1.1" 200 None ClickHouse node1 started ClickHouse node1 started Executing query DROP DATABASE IF EXISTS db on node1 Executing query DROP DATABASE IF EXISTS db on node1 Executing query DROP DATABASE IF EXISTS db2 on node1 Executing query DROP DATABASE IF EXISTS db2 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query DROP DATABASE IF EXISTS db3 on node1 Executing query DROP DATABASE IF EXISTS db3 on node1 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '1' on node1 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '1' on node1 Executing query CREATE DATABASE db UUID '10000000-1000-4000-8000-000000000001' on node1 Executing query CREATE DATABASE db UUID '10000000-1000-4000-8000-000000000001' on node1 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query CREATE TABLE db.log UUID '10000000-1000-4000-8000-000000000002' ENGINE=Log AS SELECT 1 on node1 Executing query CREATE TABLE db.log UUID '10000000-1000-4000-8000-000000000002' ENGINE=Log AS SELECT 1 on node1 Executing query CREATE TABLE db.mt UUID '10000000-1000-4000-8000-000000000003' ENGINE=MergeTree ORDER BY tuple() AS SELECT 1 on node1 Executing query CREATE TABLE db.mt UUID '10000000-1000-4000-8000-000000000003' ENGINE=MergeTree ORDER BY tuple() AS SELECT 1 on node1 Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query CREATE TABLE db.mem UUID '10000000-1000-4000-8000-000000000004' ENGINE=Memory AS SELECT 1 on node1 Executing query CREATE TABLE db.mem UUID '10000000-1000-4000-8000-000000000004' ENGINE=Memory AS SELECT 1 on node1 Executing query CREATE DATABASE db2 UUID '20000000-1000-4000-8000-000000000001' on node1 Executing query CREATE DATABASE db2 UUID '20000000-1000-4000-8000-000000000001' on node1 Executing query CREATE TABLE db2.log UUID '20000000-1000-4000-8000-000000000002' ENGINE=Log AS SELECT 1 on node1 Executing query CREATE TABLE db2.log UUID '20000000-1000-4000-8000-000000000002' ENGINE=Log AS SELECT 1 on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DETACH DATABASE db2 on node1 Executing query DETACH DATABASE db2 on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query CREATE DATABASE db3 UUID '30000000-1000-4000-8000-000000000001' on node1 Executing query CREATE DATABASE db3 UUID '30000000-1000-4000-8000-000000000001' on node1 No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/53f318a8ec1b0ab43d1f5437da994f99877f8f98b92e189127f88e2789040b24/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/53f318a8ec1b0ab43d1f5437da994f99877f8f98b92e189127f88e2789040b24/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/53f318a8ec1b0ab43d1f5437da994f99877f8f98b92e189127f88e2789040b24/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/53f318a8ec1b0ab43d1f5437da994f99877f8f98b92e189127f88e2789040b24/json HTTP/1.1" 200 586 Executing query CREATE TABLE db3.log UUID '30000000-1000-4000-8000-000000000002' ENGINE=Log AS SELECT 1 on node1 Executing query CREATE TABLE db3.log UUID '30000000-1000-4000-8000-000000000002' ENGINE=Log AS SELECT 1 on node1 Executing query CREATE TABLE db3.log2 UUID '30000000-1000-4000-8000-000000000003' ENGINE=Log AS SELECT 1 on node1 Executing query CREATE TABLE db3.log2 UUID '30000000-1000-4000-8000-000000000003' ENGINE=Log AS SELECT 1 on node1 Executing query DETACH TABLE db3.log on node1 Executing query DETACH TABLE db3.log on node1 Executing query DETACH TABLE db3.log2 PERMANENTLY on node1 Executing query DETACH TABLE db3.log2 PERMANENTLY on node1 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store] Stdout:total 16 Stdout:total 16 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 0e7 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 0e7 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 100 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 100 Stdout:drwxr-x--- 4 root root 4096 Apr 2 04:15 200 Stdout:drwxr-x--- 4 root root 4096 Apr 2 04:15 200 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 300 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 300 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/100'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/100'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/100] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/100] Stdout:total 12 Stdout:total 12 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 4 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000003 Stdout:drwxr-x--- 4 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000003 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/200'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/200'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/200] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/200] Stdout:total 8 Stdout:total 8 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000002 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/300'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/300'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/300] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/300] Stdout:total 12 Stdout:total 12 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root rootteststorecleanup-gw1-node1-1 bash -c ps -C clickhouse] Command:[docker exec -u root rootteststorecleanup-gw1-node1-1 bash -c ps -C clickhouse] Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:00 clickhouse Stdout: 10 ? 00:00:00 clickhouse run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill -9 clickhouse'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill -9 clickhouse'] Command:[docker exec -u root rootteststorecleanup-gw1-node1-1 bash -c pkill -9 clickhouse] Command:[docker exec -u root rootteststorecleanup-gw1-node1-1 bash -c pkill -9 clickhouse] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:10 Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4031 Stdout:4031 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4031 Stdout:4031 Executing query select 20 on node_default Executing query select 20 on node_default Executing query SYSTEM FLUSH LOGS on node_default Executing query SYSTEM FLUSH LOGS on node_default Executing query SHOW CREATE TABLE system.query_log FORMAT TSVRaw on node_default Executing query SHOW CREATE TABLE system.query_log FORMAT TSVRaw on node_default With storage policy, SHOW CREATE TABLE system.query_log is: CREATE TABLE system.query_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `type` Enum8('QueryStart' = 1, 'QueryFinish' = 2, 'ExceptionBeforeStart' = 3, 'ExceptionWhileProcessing' = 4) COMMENT 'Type of an event that occurred when executing the query.', `event_date` Date COMMENT 'Query starting date.', `event_time` DateTime COMMENT 'Query starting time.', `event_time_microseconds` DateTime64(6) COMMENT 'Query starting time with microseconds precision.', `query_start_time` DateTime COMMENT 'Start time of query execution.', `query_start_time_microseconds` DateTime64(6) COMMENT 'Start time of query execution with microsecond precision.', `query_duration_ms` UInt64 COMMENT 'Duration of query execution in milliseconds.', `read_rows` UInt64 COMMENT 'Total number of rows read from all tables and table functions participated in query. It includes usual subqueries, subqueries for IN and JOIN. For distributed queries read_rows includes the total number of rows read at all replicas. Each replica sends it\'s read_rows value, and the server-initiator of the query summarizes all received and local values. The cache volumes do not affect this value.', `read_bytes` UInt64 COMMENT 'Total number of bytes read from all tables and table functions participated in query. It includes usual subqueries, subqueries for IN and JOIN. For distributed queries read_bytes includes the total number of rows read at all replicas. Each replica sends it\'s read_bytes value, and the server-initiator of the query summarizes all received and local values. The cache volumes do not affect this value.', `written_rows` UInt64 COMMENT 'For INSERT queries, the number of written rows. For other queries, the column value is 0.', `written_bytes` UInt64 COMMENT 'For INSERT queries, the number of written bytes (uncompressed). For other queries, the column value is 0.', `result_rows` UInt64 COMMENT 'Number of rows in a result of the SELECT query, or a number of rows in the INSERT query.', `result_bytes` UInt64 COMMENT 'RAM volume in bytes used to store a query result.', `memory_usage` UInt64 COMMENT 'Memory consumption by the query.', `current_database` LowCardinality(String) COMMENT 'Name of the current database.', `query` String COMMENT ' Query string.', `formatted_query` String COMMENT 'Formatted query string.', `normalized_query_hash` UInt64 COMMENT 'A numeric hash value, such as it is identical for queries differ only by values of literals.', `query_kind` LowCardinality(String) COMMENT 'Type of the query.', `databases` Array(LowCardinality(String)) COMMENT 'Names of the databases present in the query.', `tables` Array(LowCardinality(String)) COMMENT 'Names of the tables present in the query.', `columns` Array(LowCardinality(String)) COMMENT 'Names of the columns present in the query.', `partitions` Array(LowCardinality(String)) COMMENT 'Names of the partitions present in the query.', `projections` Array(LowCardinality(String)) COMMENT 'Names of the projections used during the query execution.', `views` Array(LowCardinality(String)) COMMENT 'Names of the (materialized or live) views present in the query.', `exception_code` Int32 COMMENT 'Code of an exception.', `exception` String COMMENT 'Exception message.', `stack_trace` String COMMENT 'Stack trace. An empty string, if the query was completed successfully.', `is_initial_query` UInt8 COMMENT 'Query type. Possible values: 1 — query was initiated by the client, 0 — query was initiated by another query as part of distributed query execution.', `user` LowCardinality(String) COMMENT 'Name of the user who initiated the current query.', `query_id` String COMMENT 'ID of the query.', `address` IPv6 COMMENT 'IP address that was used to make the query.', `port` UInt16 COMMENT 'The client port that was used to make the query.', `initial_user` LowCardinality(String) COMMENT 'Name of the user who ran the initial query (for distributed query execution).', `initial_query_id` String COMMENT 'ID of the initial query (for distributed query execution).', `initial_address` IPv6 COMMENT 'IP address that the parent query was launched from.', `initial_port` UInt16 COMMENT 'The client port that was used to make the parent query.', `initial_query_start_time` DateTime COMMENT 'Initial query starting time (for distributed query execution).', `initial_query_start_time_microseconds` DateTime64(6) COMMENT 'Initial query starting time with microseconds precision (for distributed query execution).', `interface` UInt8 COMMENT 'Interface that the query was initiated from. Possible values: 1 — TCP, 2 — HTTP.', `is_secure` UInt8 COMMENT 'The flag whether a query was executed over a secure interface', `os_user` LowCardinality(String) COMMENT 'Operating system username who runs clickhouse-client.', `client_hostname` LowCardinality(String) COMMENT 'Hostname of the client machine where the clickhouse-client or another TCP client is run.', `client_name` LowCardinality(String) COMMENT 'The clickhouse-client or another TCP client name.', `client_revision` UInt32 COMMENT 'Revision of the clickhouse-client or another TCP client.', `client_version_major` UInt32 COMMENT 'Major version of the clickhouse-client or another TCP client.', `client_version_minor` UInt32 COMMENT 'Minor version of the clickhouse-client or another TCP client.', `client_version_patch` UInt32 COMMENT 'Patch component of the clickhouse-client or another TCP client version.', `script_query_number` UInt32 COMMENT 'The query number in a script with multiple queries for clickhouse-client.', `script_line_number` UInt32 COMMENT 'The line number of the query start in a script with multiple queries for clickhouse-client.', `http_method` UInt8 COMMENT 'HTTP method that initiated the query. Possible values: 0 — The query was launched from the TCP interface, 1 — GET method was used, 2 — POST method was used.', `http_user_agent` LowCardinality(String) COMMENT 'HTTP header UserAgent passed in the HTTP query.', `http_referer` String COMMENT 'HTTP header Referer passed in the HTTP query (contains an absolute or partial address of the page making the query).', `forwarded_for` String COMMENT 'HTTP header X-Forwarded-For passed in the HTTP query.', `quota_key` String COMMENT 'The quota key specified in the quotas setting (see keyed).', `distributed_depth` UInt64 COMMENT 'How many times a query was forwarded between servers.', `revision` UInt32 COMMENT 'ClickHouse revision.', `log_comment` String COMMENT 'Log comment. It can be set to arbitrary string no longer than max_query_size. An empty string if it is not defined.', `thread_ids` Array(UInt64) COMMENT 'Thread ids that are participating in query execution. These threads may not have run simultaneously.', `peak_threads_usage` UInt64 COMMENT 'Maximum count of simultaneous threads executing the query.', `ProfileEvents` Map(LowCardinality(String), UInt64) COMMENT 'ProfileEvents that measure different metrics. The description of them could be found in the table system.events', `Settings` Map(LowCardinality(String), LowCardinality(String)) COMMENT 'Settings that were changed when the client ran the query. To enable logging changes to settings, set the log_query_settings parameter to 1.', `used_aggregate_functions` Array(LowCardinality(String)) COMMENT 'Canonical names of aggregate functions, which were used during query execution.', `used_aggregate_function_combinators` Array(LowCardinality(String)) COMMENT 'Canonical names of aggregate functions combinators, which were used during query execution.', `used_database_engines` Array(LowCardinality(String)) COMMENT 'Canonical names of database engines, which were used during query execution.', `used_data_type_families` Array(LowCardinality(String)) COMMENT 'Canonical names of data type families, which were used during query execution.', `used_dictionaries` Array(LowCardinality(String)) COMMENT 'Canonical names of dictionaries, which were used during query execution.', `used_formats` Array(LowCardinality(String)) COMMENT 'Canonical names of formats, which were used during query execution.', `used_functions` Array(LowCardinality(String)) COMMENT 'Canonical names of functions, which were used during query execution.', `used_storages` Array(LowCardinality(String)) COMMENT 'Canonical names of storages, which were used during query execution.', `used_table_functions` Array(LowCardinality(String)) COMMENT 'Canonical names of table functions, which were used during query execution.', `used_row_policies` Array(LowCardinality(String)) COMMENT 'The list of row policies names that were used during query execution.', `used_privileges` Array(LowCardinality(String)) COMMENT 'Privileges which were successfully checked during query execution.', `missing_privileges` Array(LowCardinality(String)) COMMENT 'Privileges that are missing during query execution.', `transaction_id` Tuple( UInt64, UInt64, UUID) COMMENT 'The identifier of the transaction in scope of which this query was executed.', `query_cache_usage` Enum8('Unknown' = 0, 'None' = 1, 'Write' = 2, 'Read' = 3) COMMENT 'Usage of the query cache during query execution. Values: \'Unknown\' = Status unknown, \'None\' = The query result was neither written into nor read from the query cache, \'Write\' = The query result was written into the query cache, \'Read\' = The query result was read from the query cache.', `asynchronous_read_counters` Map(LowCardinality(String), UInt64) COMMENT 'Metrics for asynchronous reading.', `ProfileEvents.Names` Array(LowCardinality(String)) ALIAS mapKeys(ProfileEvents), `ProfileEvents.Values` Array(UInt64) ALIAS mapValues(ProfileEvents), `Settings.Names` Array(LowCardinality(String)) ALIAS mapKeys(Settings), `Settings.Values` Array(LowCardinality(String)) ALIAS mapValues(Settings) ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains information about executed queries, for example, start time, duration of processing, error messages.\n\nIt is safe to truncate or drop this table at any time.' With storage policy, SHOW CREATE TABLE system.query_log is: CREATE TABLE system.query_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `type` Enum8('QueryStart' = 1, 'QueryFinish' = 2, 'ExceptionBeforeStart' = 3, 'ExceptionWhileProcessing' = 4) COMMENT 'Type of an event that occurred when executing the query.', `event_date` Date COMMENT 'Query starting date.', `event_time` DateTime COMMENT 'Query starting time.', `event_time_microseconds` DateTime64(6) COMMENT 'Query starting time with microseconds precision.', `query_start_time` DateTime COMMENT 'Start time of query execution.', `query_start_time_microseconds` DateTime64(6) COMMENT 'Start time of query execution with microsecond precision.', `query_duration_ms` UInt64 COMMENT 'Duration of query execution in milliseconds.', `read_rows` UInt64 COMMENT 'Total number of rows read from all tables and table functions participated in query. It includes usual subqueries, subqueries for IN and JOIN. For distributed queries read_rows includes the total number of rows read at all replicas. Each replica sends it\'s read_rows value, and the server-initiator of the query summarizes all received and local values. The cache volumes do not affect this value.', `read_bytes` UInt64 COMMENT 'Total number of bytes read from all tables and table functions participated in query. It includes usual subqueries, subqueries for IN and JOIN. For distributed queries read_bytes includes the total number of rows read at all replicas. Each replica sends it\'s read_bytes value, and the server-initiator of the query summarizes all received and local values. The cache volumes do not affect this value.', `written_rows` UInt64 COMMENT 'For INSERT queries, the number of written rows. For other queries, the column value is 0.', `written_bytes` UInt64 COMMENT 'For INSERT queries, the number of written bytes (uncompressed). For other queries, the column value is 0.', `result_rows` UInt64 COMMENT 'Number of rows in a result of the SELECT query, or a number of rows in the INSERT query.', `result_bytes` UInt64 COMMENT 'RAM volume in bytes used to store a query result.', `memory_usage` UInt64 COMMENT 'Memory consumption by the query.', `current_database` LowCardinality(String) COMMENT 'Name of the current database.', `query` String COMMENT ' Query string.', `formatted_query` String COMMENT 'Formatted query string.', `normalized_query_hash` UInt64 COMMENT 'A numeric hash value, such as it is identical for queries differ only by values of literals.', `query_kind` LowCardinality(String) COMMENT 'Type of the query.', `databases` Array(LowCardinality(String)) COMMENT 'Names of the databases present in the query.', `tables` Array(LowCardinality(String)) COMMENT 'Names of the tables present in the query.', `columns` Array(LowCardinality(String)) COMMENT 'Names of the columns present in the query.', `partitions` Array(LowCardinality(String)) COMMENT 'Names of the partitions present in the query.', `projections` Array(LowCardinality(String)) COMMENT 'Names of the projections used during the query execution.', `views` Array(LowCardinality(String)) COMMENT 'Names of the (materialized or live) views present in the query.', `exception_code` Int32 COMMENT 'Code of an exception.', `exception` String COMMENT 'Exception message.', `stack_trace` String COMMENT 'Stack trace. An empty string, if the query was completed successfully.', `is_initial_query` UInt8 COMMENT 'Query type. Possible values: 1 — query was initiated by the client, 0 — query was initiated by another query as part of distributed query execution.', `user` LowCardinality(String) COMMENT 'Name of the user who initiated the current query.', `query_id` String COMMENT 'ID of the query.', `address` IPv6 COMMENT 'IP address that was used to make the query.', `port` UInt16 COMMENT 'The client port that was used to make the query.', `initial_user` LowCardinality(String) COMMENT 'Name of the user who ran the initial query (for distributed query execution).', `initial_query_id` String COMMENT 'ID of the initial query (for distributed query execution).', `initial_address` IPv6 COMMENT 'IP address that the parent query was launched from.', `initial_port` UInt16 COMMENT 'The client port that was used to make the parent query.', `initial_query_start_time` DateTime COMMENT 'Initial query starting time (for distributed query execution).', `initial_query_start_time_microseconds` DateTime64(6) COMMENT 'Initial query starting time with microseconds precision (for distributed query execution).', `interface` UInt8 COMMENT 'Interface that the query was initiated from. Possible values: 1 — TCP, 2 — HTTP.', `is_secure` UInt8 COMMENT 'The flag whether a query was executed over a secure interface', `os_user` LowCardinality(String) COMMENT 'Operating system username who runs clickhouse-client.', `client_hostname` LowCardinality(String) COMMENT 'Hostname of the client machine where the clickhouse-client or another TCP client is run.', `client_name` LowCardinality(String) COMMENT 'The clickhouse-client or another TCP client name.', `client_revision` UInt32 COMMENT 'Revision of the clickhouse-client or another TCP client.', `client_version_major` UInt32 COMMENT 'Major version of the clickhouse-client or another TCP client.', `client_version_minor` UInt32 COMMENT 'Minor version of the clickhouse-client or another TCP client.', `client_version_patch` UInt32 COMMENT 'Patch component of the clickhouse-client or another TCP client version.', `script_query_number` UInt32 COMMENT 'The query number in a script with multiple queries for clickhouse-client.', `script_line_number` UInt32 COMMENT 'The line number of the query start in a script with multiple queries for clickhouse-client.', `http_method` UInt8 COMMENT 'HTTP method that initiated the query. Possible values: 0 — The query was launched from the TCP interface, 1 — GET method was used, 2 — POST method was used.', `http_user_agent` LowCardinality(String) COMMENT 'HTTP header UserAgent passed in the HTTP query.', `http_referer` String COMMENT 'HTTP header Referer passed in the HTTP query (contains an absolute or partial address of the page making the query).', `forwarded_for` String COMMENT 'HTTP header X-Forwarded-For passed in the HTTP query.', `quota_key` String COMMENT 'The quota key specified in the quotas setting (see keyed).', `distributed_depth` UInt64 COMMENT 'How many times a query was forwarded between servers.', `revision` UInt32 COMMENT 'ClickHouse revision.', `log_comment` String COMMENT 'Log comment. It can be set to arbitrary string no longer than max_query_size. An empty string if it is not defined.', `thread_ids` Array(UInt64) COMMENT 'Thread ids that are participating in query execution. These threads may not have run simultaneously.', `peak_threads_usage` UInt64 COMMENT 'Maximum count of simultaneous threads executing the query.', `ProfileEvents` Map(LowCardinality(String), UInt64) COMMENT 'ProfileEvents that measure different metrics. The description of them could be found in the table system.events', `Settings` Map(LowCardinality(String), LowCardinality(String)) COMMENT 'Settings that were changed when the client ran the query. To enable logging changes to settings, set the log_query_settings parameter to 1.', `used_aggregate_functions` Array(LowCardinality(String)) COMMENT 'Canonical names of aggregate functions, which were used during query execution.', `used_aggregate_function_combinators` Array(LowCardinality(String)) COMMENT 'Canonical names of aggregate functions combinators, which were used during query execution.', `used_database_engines` Array(LowCardinality(String)) COMMENT 'Canonical names of database engines, which were used during query execution.', `used_data_type_families` Array(LowCardinality(String)) COMMENT 'Canonical names of data type families, which were used during query execution.', `used_dictionaries` Array(LowCardinality(String)) COMMENT 'Canonical names of dictionaries, which were used during query execution.', `used_formats` Array(LowCardinality(String)) COMMENT 'Canonical names of formats, which were used during query execution.', `used_functions` Array(LowCardinality(String)) COMMENT 'Canonical names of functions, which were used during query execution.', `used_storages` Array(LowCardinality(String)) COMMENT 'Canonical names of storages, which were used during query execution.', `used_table_functions` Array(LowCardinality(String)) COMMENT 'Canonical names of table functions, which were used during query execution.', `used_row_policies` Array(LowCardinality(String)) COMMENT 'The list of row policies names that were used during query execution.', `used_privileges` Array(LowCardinality(String)) COMMENT 'Privileges which were successfully checked during query execution.', `missing_privileges` Array(LowCardinality(String)) COMMENT 'Privileges that are missing during query execution.', `transaction_id` Tuple( UInt64, UInt64, UUID) COMMENT 'The identifier of the transaction in scope of which this query was executed.', `query_cache_usage` Enum8('Unknown' = 0, 'None' = 1, 'Write' = 2, 'Read' = 3) COMMENT 'Usage of the query cache during query execution. Values: \'Unknown\' = Status unknown, \'None\' = The query result was neither written into nor read from the query cache, \'Write\' = The query result was written into the query cache, \'Read\' = The query result was read from the query cache.', `asynchronous_read_counters` Map(LowCardinality(String), UInt64) COMMENT 'Metrics for asynchronous reading.', `ProfileEvents.Names` Array(LowCardinality(String)) ALIAS mapKeys(ProfileEvents), `ProfileEvents.Values` Array(UInt64) ALIAS mapValues(ProfileEvents), `Settings.Names` Array(LowCardinality(String)) ALIAS mapKeys(Settings), `Settings.Values` Array(LowCardinality(String)) ALIAS mapValues(Settings) ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains information about executed queries, for example, start time, duration of processing, error messages.\n\nIt is safe to truncate or drop this table at any time.' Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query SHOW CREATE TABLE system.query_metric_log FORMAT TSVRaw on node_default Executing query SHOW CREATE TABLE system.query_metric_log FORMAT TSVRaw on node_default With storage policy, SHOW CREATE TABLE system.query_metric_log is: CREATE TABLE system.query_metric_log ( `query_id` String COMMENT 'Query ID.' CODEC(ZSTD(1)), `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.' CODEC(ZSTD(1)), `event_date` Date COMMENT 'Event date.' CODEC(Delta(2), ZSTD(1)), `event_time` DateTime COMMENT 'Event time.' CODEC(Delta(4), ZSTD(1)), `event_time_microseconds` DateTime64(6) COMMENT 'Event time with microseconds resolution.' CODEC(Delta(4), ZSTD(1)), `memory_usage` UInt64 COMMENT 'Amount of RAM the query uses. It might not include some types of dedicated memory.', `peak_memory_usage` UInt64 COMMENT 'Maximum amount of RAM the query used.', `ProfileEvent_Query` UInt64 COMMENT 'Number of queries to be interpreted and potentially executed. Does not include queries that failed to parse or were rejected due to AST size limits, quota limits or limits on the number of simultaneously running queries. May include internal queries initiated by ClickHouse itself. Does not count subqueries.', `ProfileEvent_SelectQuery` UInt64 COMMENT 'Same as Query, but only for SELECT queries.', `ProfileEvent_InsertQuery` UInt64 COMMENT 'Same as Query, but only for INSERT queries.', `ProfileEvent_InitialQuery` UInt64 COMMENT 'Same as Query, but only counts initial queries (see is_initial_query).', `ProfileEvent_QueriesWithSubqueries` UInt64 COMMENT 'Count queries with all subqueries', `ProfileEvent_SelectQueriesWithSubqueries` UInt64 COMMENT 'Count SELECT queries with all subqueries', `ProfileEvent_InsertQueriesWithSubqueries` UInt64 COMMENT 'Count INSERT queries with all subqueries', `ProfileEvent_SelectQueriesWithPrimaryKeyUsage` UInt64 COMMENT 'Count SELECT queries which use the primary key to evaluate the WHERE condition', `ProfileEvent_AsyncInsertQuery` UInt64 COMMENT 'Same as InsertQuery, but only for asynchronous INSERT queries.', `ProfileEvent_AsyncInsertBytes` UInt64 COMMENT 'Data size in bytes of asynchronous INSERT queries.', `ProfileEvent_AsyncInsertRows` UInt64 COMMENT 'Number of rows inserted by asynchronous INSERT queries.', `ProfileEvent_AsyncInsertCacheHits` UInt64 COMMENT 'Number of times a duplicate hash id has been found in asynchronous INSERT hash id cache.', `ProfileEvent_FailedQuery` UInt64 COMMENT 'Number of failed queries.', `ProfileEvent_FailedSelectQuery` UInt64 COMMENT 'Same as FailedQuery, but only for SELECT queries.', `ProfileEvent_FailedInsertQuery` UInt64 COMMENT 'Same as FailedQuery, but only for INSERT queries.', `ProfileEvent_FailedAsyncInsertQuery` UInt64 COMMENT 'Number of failed ASYNC INSERT queries.', `ProfileEvent_QueryTimeMicroseconds` UInt64 COMMENT 'Total time of all queries.', `ProfileEvent_SelectQueryTimeMicroseconds` UInt64 COMMENT 'Total time of SELECT queries.', `ProfileEvent_InsertQueryTimeMicroseconds` UInt64 COMMENT 'Total time of INSERT queries.', `ProfileEvent_OtherQueryTimeMicroseconds` UInt64 COMMENT 'Total time of queries that are not SELECT or INSERT.', `ProfileEvent_FileOpen` UInt64 COMMENT 'Number of files opened.', `ProfileEvent_Seek` UInt64 COMMENT 'Number of times the \'lseek\' function was called.', `ProfileEvent_ReadBufferFromFileDescriptorRead` UInt64 COMMENT 'Number of reads (read/pread) from a file descriptor. Does not include sockets.', `ProfileEvent_ReadBufferFromFileDescriptorReadFailed` UInt64 COMMENT 'Number of times the read (read/pread) from a file descriptor have failed.', `ProfileEvent_ReadBufferFromFileDescriptorReadBytes` UInt64 COMMENT 'Number of bytes read from file descriptors. If the file is compressed, this will show the compressed data size.', `ProfileEvent_WriteBufferFromFileDescriptorWrite` UInt64 COMMENT 'Number of writes (write/pwrite) to a file descriptor. Does not include sockets.', `ProfileEvent_WriteBufferFromFileDescriptorWriteFailed` UInt64 COMMENT 'Number of times the write (write/pwrite) to a file descriptor have failed.', `ProfileEvent_WriteBufferFromFileDescriptorWriteBytes` UInt64 COMMENT 'Number of bytes written to file descriptors. If the file is compressed, this will show compressed data size.', `ProfileEvent_FileSync` UInt64 COMMENT 'Number of times the F_FULLFSYNC/fsync/fdatasync function was called for files.', `ProfileEvent_DirectorySync` UInt64 COMMENT 'Number of times the F_FULLFSYNC/fsync/fdatasync function was called for directories.', `ProfileEvent_FileSyncElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for F_FULLFSYNC/fsync/fdatasync syscall for files.', `ProfileEvent_DirectorySyncElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for F_FULLFSYNC/fsync/fdatasync syscall for directories.', `ProfileEvent_ReadCompressedBytes` UInt64 COMMENT 'Number of bytes (the number of bytes before decompression) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferBlocks` UInt64 COMMENT 'Number of compressed blocks (the blocks of data that are compressed independent of each other) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferBytes` UInt64 COMMENT 'Number of uncompressed bytes (the number of bytes after decompression) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferChecksumDoesntMatch` UInt64 COMMENT 'Number of times the compressed block checksum did not match.', `ProfileEvent_CompressedReadBufferChecksumDoesntMatchSingleBitMismatch` UInt64 COMMENT 'Number of times a compressed block checksum mismatch was caused by a single-bit difference.', `ProfileEvent_CompressedReadBufferChecksumDoesntMatchMicroseconds` UInt64 COMMENT 'Total time spent detecting bit-flips due to compressed block checksum mismatches.', `ProfileEvent_UncompressedCacheHits` UInt64 COMMENT 'Number of times a block of data has been found in the uncompressed cache (and decompression was avoided).', `ProfileEvent_UncompressedCacheMisses` UInt64 COMMENT 'Number of times a block of data has not been found in the uncompressed cache (and required decompression).', `ProfileEvent_UncompressedCacheWeightLost` UInt64 COMMENT 'Number of bytes evicted from the uncompressed cache.', `ProfileEvent_MMappedFileCacheHits` UInt64 COMMENT 'Number of times a file has been found in the MMap cache (for the \'mmap\' read_method), so we didn\'t have to mmap it again.', `ProfileEvent_MMappedFileCacheMisses` UInt64 COMMENT 'Number of times a file has not been found in the MMap cache (for the \'mmap\' read_method), so we had to mmap it again.', `ProfileEvent_OpenedFileCacheHits` UInt64 COMMENT 'Number of times a file has been found in the opened file cache, so we didn\'t have to open it again.', `ProfileEvent_OpenedFileCacheMisses` UInt64 COMMENT 'Number of times a file has been found in the opened file cache, so we had to open it again.', `ProfileEvent_OpenedFileCacheMicroseconds` UInt64 COMMENT 'Amount of time spent executing OpenedFileCache methods.', `ProfileEvent_AIOWrite` UInt64 COMMENT 'Number of writes with Linux or FreeBSD AIO interface', `ProfileEvent_AIOWriteBytes` UInt64 COMMENT 'Number of bytes written with Linux or FreeBSD AIO interface', `ProfileEvent_AIORead` UInt64 COMMENT 'Number of reads with Linux or FreeBSD AIO interface', `ProfileEvent_AIOReadBytes` UInt64 COMMENT 'Number of bytes read with Linux or FreeBSD AIO interface', `ProfileEvent_IOBufferAllocs` UInt64 COMMENT 'Number of allocations of IO buffers (for ReadBuffer/WriteBuffer).', `ProfileEvent_IOBufferAllocBytes` UInt64 COMMENT 'Number of bytes allocated for IO buffers (for ReadBuffer/WriteBuffer).', `ProfileEvent_ArenaAllocChunks` UInt64 COMMENT 'Number of chunks allocated for memory Arena (used for GROUP BY and similar operations)', `ProfileEvent_ArenaAllocBytes` UInt64 COMMENT 'Number of bytes allocated for memory Arena (used for GROUP BY and similar operations)', `ProfileEvent_FunctionExecute` UInt64 COMMENT 'Number of SQL ordinary function calls (SQL functions are called on per-block basis, so this number represents the number of blocks).', `ProfileEvent_TableFunctionExecute` UInt64 COMMENT 'Number of table function calls.', `ProfileEvent_DefaultImplementationForNullsRows` UInt64 COMMENT 'Number of rows processed by default implementation for nulls in function execution', `ProfileEvent_DefaultImplementationForNullsRowsWithNulls` UInt64 COMMENT 'Number of rows which contain null values processed by default implementation for nulls in function execution', `ProfileEvent_MarkCacheHits` UInt64 COMMENT 'Number of times an entry has been found in the mark cache, so we didn\'t have to load a mark file.', `ProfileEvent_MarkCacheMisses` UInt64 COMMENT 'Number of times an entry has not been found in the mark cache, so we had to load a mark file in memory, which is a costly operation, adding to query latency.', `ProfileEvent_PrimaryIndexCacheHits` UInt64 COMMENT 'Number of times an entry has been found in the primary index cache, so we didn\'t have to load a index file.', `ProfileEvent_PrimaryIndexCacheMisses` UInt64 COMMENT 'Number of times an entry has not been found in the primary index cache, so we had to load a index file in memory, which is a costly operation, adding to query latency.', `ProfileEvent_SkippingIndexCacheHits` UInt64 COMMENT 'Number of times an index granule has been found in the skipping index cache.', `ProfileEvent_SkippingIndexCacheMisses` UInt64 COMMENT 'Number of times an index granule has not been found in the skipping index cache and had to be read from disk.', `ProfileEvent_SkippingIndexCacheWeightLost` UInt64 COMMENT 'Approximate number of bytes evicted from the secondary index cache.', `ProfileEvent_QueryCacheHits` UInt64 COMMENT 'Number of times a query result has been found in the query cache (and query computation was avoided). Only updated for SELECT queries with SETTING use_query_cache = 1.', `ProfileEvent_QueryCacheMisses` UInt64 COMMENT 'Number of times a query result has not been found in the query cache (and required query computation). Only updated for SELECT queries with SETTING use_query_cache = 1.', `ProfileEvent_PageCacheChunkMisses` UInt64 COMMENT 'Number of times a chunk has not been found in the userspace page cache.', `ProfileEvent_PageCacheChunkShared` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, already in use by another thread.', `ProfileEvent_PageCacheChunkDataHits` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, with all pages intact.', `ProfileEvent_PageCacheChunkDataPartialHits` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, but some of its pages were evicted by the OS.', `ProfileEvent_PageCacheChunkDataMisses` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, but all its pages were evicted by the OS.', `ProfileEvent_PageCacheBytesUnpinnedRoundedToPages` UInt64 COMMENT 'Total size of populated pages in chunks that became evictable in PageCache. Rounded up to whole pages.', `ProfileEvent_PageCacheBytesUnpinnedRoundedToHugePages` UInt64 COMMENT 'See PageCacheBytesUnpinnedRoundedToPages, but rounded to huge pages. Use the ratio between the two as a measure of memory waste from using huge pages.', `ProfileEvent_CreatedReadBufferOrdinary` UInt64 COMMENT 'Number of times ordinary read buffer was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferDirectIO` UInt64 COMMENT 'Number of times a read buffer with O_DIRECT was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferDirectIOFailed` UInt64 COMMENT 'Number of times a read buffer with O_DIRECT was attempted to be created for reading data (while choosing among other read methods), but the OS did not allow it (due to lack of filesystem support or other reasons) and we fallen back to the ordinary reading method.', `ProfileEvent_CreatedReadBufferMMap` UInt64 COMMENT 'Number of times a read buffer using \'mmap\' was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferMMapFailed` UInt64 COMMENT 'Number of times a read buffer with \'mmap\' was attempted to be created for reading data (while choosing among other read methods), but the OS did not allow it (due to lack of filesystem support or other reasons) and we fallen back to the ordinary reading method.', `ProfileEvent_DiskReadElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for read syscall. This include reads from page cache.', `ProfileEvent_DiskWriteElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for write syscall. This include writes to page cache.', `ProfileEvent_NetworkReceiveElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for data to receive or receiving data from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkSendElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for data to send to network or sending data to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkReceiveBytes` UInt64 COMMENT 'Total number of bytes received from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkSendBytes` UInt64 COMMENT 'Total number of bytes send to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_GlobalThreadPoolExpansions` UInt64 COMMENT 'Counts the total number of times new threads have been added to the global thread pool. This metric indicates the frequency of expansions in the global thread pool to accommodate increased processing demands.', `ProfileEvent_GlobalThreadPoolShrinks` UInt64 COMMENT 'Counts the total number of times the global thread pool has shrunk by removing threads. This occurs when the number of idle threads exceeds max_thread_pool_free_size, indicating adjustments in the global thread pool size in response to decreased thread utilization.', `ProfileEvent_GlobalThreadPoolThreadCreationMicroseconds` UInt64 COMMENT 'Total time spent waiting for new threads to start.', `ProfileEvent_GlobalThreadPoolLockWaitMicroseconds` UInt64 COMMENT 'Total time threads have spent waiting for locks in the global thread pool.', `ProfileEvent_GlobalThreadPoolJobs` UInt64 COMMENT 'Counts the number of jobs that have been pushed to the global thread pool.', `ProfileEvent_GlobalThreadPoolJobWaitTimeMicroseconds` UInt64 COMMENT 'Measures the elapsed time from when a job is scheduled in the thread pool to when it is picked up for execution by a worker thread. This metric helps identify delays in job processing, indicating the responsiveness of the thread pool to new tasks.', `ProfileEvent_LocalThreadPoolExpansions` UInt64 COMMENT 'Counts the total number of times threads have been borrowed from the global thread pool to expand local thread pools.', `ProfileEvent_LocalThreadPoolShrinks` UInt64 COMMENT 'Counts the total number of times threads have been returned to the global thread pool from local thread pools.', `ProfileEvent_LocalThreadPoolThreadCreationMicroseconds` UInt64 COMMENT 'Total time local thread pools have spent waiting to borrow a thread from the global pool.', `ProfileEvent_LocalThreadPoolLockWaitMicroseconds` UInt64 COMMENT 'Total time threads have spent waiting for locks in the local thread pools.', `ProfileEvent_LocalThreadPoolJobs` UInt64 COMMENT 'Counts the number of jobs that have been pushed to the local thread pools.', `ProfileEvent_LocalThreadPoolBusyMicroseconds` UInt64 COMMENT 'Total time threads have spent executing the actual work.', `ProfileEvent_LocalThreadPoolJobWaitTimeMicroseconds` UInt64 COMMENT 'Measures the elapsed time from when a job is scheduled in the thread pool to when it is picked up for execution by a worker thread. This metric helps identify delays in job processing, indicating the responsiveness of the thread pool to new tasks.', `ProfileEvent_DiskS3GetRequestThrottlerCount` UInt64 COMMENT 'Number of DiskS3 GET and SELECT requests passed through throttler.', `ProfileEvent_DiskS3GetRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform DiskS3 GET and SELECT request throttling.', `ProfileEvent_DiskS3PutRequestThrottlerCount` UInt64 COMMENT 'Number of DiskS3 PUT, COPY, POST and LIST requests passed through throttler.', `ProfileEvent_DiskS3PutRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform DiskS3 PUT, COPY, POST and LIST request throttling.', `ProfileEvent_S3GetRequestThrottlerCount` UInt64 COMMENT 'Number of S3 GET and SELECT requests passed through throttler.', `ProfileEvent_S3GetRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform S3 GET and SELECT request throttling.', `ProfileEvent_S3PutRequestThrottlerCount` UInt64 COMMENT 'Number of S3 PUT, COPY, POST and LIST requests passed through throttler.', `ProfileEvent_S3PutRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform S3 PUT, COPY, POST and LIST request throttling.', `ProfileEvent_RemoteReadThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_remote_read_network_bandwidth_for_server\'/\'max_remote_read_network_bandwidth\' throttler.', `ProfileEvent_RemoteReadThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_remote_read_network_bandwidth_for_server\'/\'max_remote_read_network_bandwidth\' throttling.', `ProfileEvent_RemoteWriteThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_remote_write_network_bandwidth_for_server\'/\'max_remote_write_network_bandwidth\' throttler.', `ProfileEvent_RemoteWriteThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_remote_write_network_bandwidth_for_server\'/\'max_remote_write_network_bandwidth\' throttling.', `ProfileEvent_LocalReadThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_local_read_bandwidth_for_server\'/\'max_local_read_bandwidth\' throttler.', `ProfileEvent_LocalReadThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_local_read_bandwidth_for_server\'/\'max_local_read_bandwidth\' throttling.', `ProfileEvent_LocalWriteThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_local_write_bandwidth_for_server\'/\'max_local_write_bandwidth\' throttler.', `ProfileEvent_LocalWriteThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_local_write_bandwidth_for_server\'/\'max_local_write_bandwidth\' throttling.', `ProfileEvent_ThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform all throttling settings.', `ProfileEvent_ReadTasksWithAppliedMutationsOnFly` UInt64 COMMENT 'Total number of parts for which there was any mutation applied on fly', `ProfileEvent_MutationsAppliedOnFlyInAllReadTasks` UInt64 COMMENT 'The sum of number of applied mutations on-fly for part among all read parts', `ProfileEvent_SchedulerIOReadRequests` UInt64 COMMENT 'Resource requests passed through scheduler for IO reads.', `ProfileEvent_SchedulerIOReadBytes` UInt64 COMMENT 'Bytes passed through scheduler for IO reads.', `ProfileEvent_SchedulerIOReadWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting on resource requests for IO reads.', `ProfileEvent_SchedulerIOWriteRequests` UInt64 COMMENT 'Resource requests passed through scheduler for IO writes.', `ProfileEvent_SchedulerIOWriteBytes` UInt64 COMMENT 'Bytes passed through scheduler for IO writes.', `ProfileEvent_SchedulerIOWriteWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting on resource requests for IO writes.', `ProfileEvent_QueryMaskingRulesMatch` UInt64 COMMENT 'Number of times query masking rules was successfully matched.', `ProfileEvent_ReplicatedPartFetches` UInt64 COMMENT 'Number of times a data part was downloaded from replica of a ReplicatedMergeTree table.', `ProfileEvent_ReplicatedPartFailedFetches` UInt64 COMMENT 'Number of times a data part was failed to download from replica of a ReplicatedMergeTree table.', `ProfileEvent_ObsoleteReplicatedParts` UInt64 COMMENT 'Number of times a data part was covered by another data part that has been fetched from a replica (so, we have marked a covered data part as obsolete and no longer needed).', `ProfileEvent_ReplicatedPartMerges` UInt64 COMMENT 'Number of times data parts of ReplicatedMergeTree tables were successfully merged.', `ProfileEvent_ReplicatedPartFetchesOfMerged` UInt64 COMMENT 'Number of times we prefer to download already merged part from replica of ReplicatedMergeTree table instead of performing a merge ourself (usually we prefer doing a merge ourself to save network traffic). This happens when we have not all source parts to perform a merge or when the data part is old enough.', `ProfileEvent_ReplicatedPartMutations` UInt64 COMMENT 'Number of times data parts of ReplicatedMergeTree tables were successfully mutated.', `ProfileEvent_ReplicatedPartChecks` UInt64 COMMENT 'Number of times we had to perform advanced search for a data part on replicas or to clarify the need of an existing data part.', `ProfileEvent_ReplicatedPartChecksFailed` UInt64 COMMENT 'Number of times the advanced search for a data part on replicas did not give result or when unexpected part has been found and moved away.', `ProfileEvent_ReplicatedDataLoss` UInt64 COMMENT 'Number of times a data part that we wanted doesn\'t exist on any replica (even on replicas that are offline right now). That data parts are definitely lost. This is normal due to asynchronous replication (if quorum inserts were not enabled), when the replica on which the data part was written was failed and when it became online after fail it doesn\'t contain that data part.', `ProfileEvent_ReplicatedCoveredPartsInZooKeeperOnStart` UInt64 COMMENT 'For debugging purposes. Number of parts in ZooKeeper that have a covering part, but doesn\'t exist on disk. Checked on server start.', `ProfileEvent_InsertedRows` UInt64 COMMENT 'Number of rows INSERTed to all tables.', `ProfileEvent_InsertedBytes` UInt64 COMMENT 'Number of bytes (uncompressed; for columns as they stored in memory) INSERTed to all tables.', `ProfileEvent_DelayedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a MergeTree table was throttled due to high number of active data parts for partition.', `ProfileEvent_RejectedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a MergeTree table was rejected with \'Too many parts\' exception due to high number of active data parts for partition.', `ProfileEvent_DelayedInsertsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the INSERT of a block to a MergeTree table was throttled due to high number of active data parts for partition.', `ProfileEvent_DelayedMutations` UInt64 COMMENT 'Number of times the mutation of a MergeTree table was throttled due to high number of unfinished mutations for table.', `ProfileEvent_RejectedMutations` UInt64 COMMENT 'Number of times the mutation of a MergeTree table was rejected with \'Too many mutations\' exception due to high number of unfinished mutations for table.', `ProfileEvent_DelayedMutationsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the mutation of a MergeTree table was throttled due to high number of unfinished mutations for table.', `ProfileEvent_DistributedDelayedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a Distributed table was throttled due to high number of pending bytes.', `ProfileEvent_DistributedRejectedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a Distributed table was rejected with \'Too many bytes\' exception due to high number of pending bytes.', `ProfileEvent_DistributedDelayedInsertsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the INSERT of a block to a Distributed table was throttled due to high number of pending bytes.', `ProfileEvent_DuplicatedInsertedBlocks` UInt64 COMMENT 'Number of times the INSERTed block to a ReplicatedMergeTree table was deduplicated.', `ProfileEvent_ZooKeeperInit` UInt64 COMMENT 'Number of times connection with ZooKeeper has been established.', `ProfileEvent_ZooKeeperTransactions` UInt64 COMMENT 'Number of ZooKeeper operations, which include both read and write operations as well as multi-transactions.', `ProfileEvent_ZooKeeperList` UInt64 COMMENT 'Number of \'list\' (getChildren) requests to ZooKeeper.', `ProfileEvent_ZooKeeperCreate` UInt64 COMMENT 'Number of \'create\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperRemove` UInt64 COMMENT 'Number of \'remove\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperExists` UInt64 COMMENT 'Number of \'exists\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperGet` UInt64 COMMENT 'Number of \'get\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperSet` UInt64 COMMENT 'Number of \'set\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperMulti` UInt64 COMMENT 'Number of \'multi\' requests to ZooKeeper (compound transactions).', `ProfileEvent_ZooKeeperCheck` UInt64 COMMENT 'Number of \'check\' requests to ZooKeeper. Usually they don\'t make sense in isolation, only as part of a complex transaction.', `ProfileEvent_ZooKeeperSync` UInt64 COMMENT 'Number of \'sync\' requests to ZooKeeper. These requests are rarely needed or usable.', `ProfileEvent_ZooKeeperReconfig` UInt64 COMMENT 'Number of \'reconfig\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperClose` UInt64 COMMENT 'Number of times connection with ZooKeeper has been closed voluntary.', `ProfileEvent_ZooKeeperWatchResponse` UInt64 COMMENT 'Number of times watch notification has been received from ZooKeeper.', `ProfileEvent_ZooKeeperUserExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper related to the data (no node, bad version or similar).', `ProfileEvent_ZooKeeperHardwareExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper related to network (connection loss or similar).', `ProfileEvent_ZooKeeperOtherExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper other than ZooKeeperUserExceptions and ZooKeeperHardwareExceptions.', `ProfileEvent_ZooKeeperWaitMicroseconds` UInt64 COMMENT 'Number of microseconds spent waiting for responses from ZooKeeper after creating a request, summed across all the requesting threads.', `ProfileEvent_ZooKeeperBytesSent` UInt64 COMMENT 'Number of bytes send over network while communicating with ZooKeeper.', `ProfileEvent_ZooKeeperBytesReceived` UInt64 COMMENT 'Number of bytes received over network while communicating with ZooKeeper.', `ProfileEvent_DistributedConnectionTries` UInt64 COMMENT 'Total count of distributed connection attempts.', `ProfileEvent_DistributedConnectionUsable` UInt64 COMMENT 'Total count of successful distributed connections to a usable server (with required table, but maybe stale).', `ProfileEvent_DistributedConnectionFailTry` UInt64 COMMENT 'Total count when distributed connection fails with retry.', `ProfileEvent_DistributedConnectionMissingTable` UInt64 COMMENT 'Number of times we rejected a replica from a distributed query, because it did not contain a table needed for the query.', `ProfileEvent_DistributedConnectionStaleReplica` UInt64 COMMENT 'Number of times we rejected a replica from a distributed query, because some table needed for a query had replication lag higher than the configured threshold.', `ProfileEvent_DistributedConnectionSkipReadOnlyReplica` UInt64 COMMENT 'Number of replicas skipped during INSERT into Distributed table due to replicas being read-only', `ProfileEvent_DistributedConnectionFailAtAll` UInt64 COMMENT 'Total count when distributed connection fails after all retries finished.', `ProfileEvent_HedgedRequestsChangeReplica` UInt64 COMMENT 'Total count when timeout for changing replica expired in hedged requests.', `ProfileEvent_SuspendSendingQueryToShard` UInt64 COMMENT 'Total count when sending query to shard was suspended when async_query_sending_for_remote is enabled.', `ProfileEvent_CompileFunction` UInt64 COMMENT 'Number of times a compilation of generated LLVM code (to create fused function for complex expressions) was initiated.', `ProfileEvent_CompiledFunctionExecute` UInt64 COMMENT 'Number of times a compiled function was executed.', `ProfileEvent_CompileExpressionsMicroseconds` UInt64 COMMENT 'Total time spent for compilation of expressions to LLVM code.', `ProfileEvent_CompileExpressionsBytes` UInt64 COMMENT 'Number of bytes used for expressions compilation.', `ProfileEvent_ExecuteShellCommand` UInt64 COMMENT 'Number of shell command executions.', `ProfileEvent_ExternalProcessingCompressedBytesTotal` UInt64 COMMENT 'Number of compressed bytes written by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalProcessingUncompressedBytesTotal` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalProcessingFilesTotal` UInt64 COMMENT 'Number of files used by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalSortWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for sorting in external memory.', `ProfileEvent_ExternalSortMerge` UInt64 COMMENT 'Number of times temporary files were merged for sorting in external memory.', `ProfileEvent_ExternalSortCompressedBytes` UInt64 COMMENT 'Number of compressed bytes written for sorting in external memory.', `ProfileEvent_ExternalSortUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written for sorting in external memory.', `ProfileEvent_ExternalAggregationWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for aggregation in external memory.', `ProfileEvent_ExternalAggregationMerge` UInt64 COMMENT 'Number of times temporary files were merged for aggregation in external memory.', `ProfileEvent_ExternalAggregationCompressedBytes` UInt64 COMMENT 'Number of bytes written to disk for aggregation in external memory.', `ProfileEvent_ExternalAggregationUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written to disk for aggregation in external memory.', `ProfileEvent_ExternalJoinWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for JOIN in external memory.', `ProfileEvent_ExternalJoinMerge` UInt64 COMMENT 'Number of times temporary files were merged for JOIN in external memory.', `ProfileEvent_ExternalJoinCompressedBytes` UInt64 COMMENT 'Number of compressed bytes written for JOIN in external memory.', `ProfileEvent_ExternalJoinUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written for JOIN in external memory.', `ProfileEvent_IcebergPartitionPrunnedFiles` UInt64 COMMENT 'Number of skipped files during Iceberg partition pruning', `ProfileEvent_JoinBuildTableRowCount` UInt64 COMMENT 'Total number of rows in the build table for a JOIN operation.', `ProfileEvent_JoinProbeTableRowCount` UInt64 COMMENT 'Total number of rows in the probe table for a JOIN operation.', `ProfileEvent_JoinResultRowCount` UInt64 COMMENT 'Total number of rows in the result of a JOIN operation.', `ProfileEvent_SlowRead` UInt64 COMMENT 'Number of reads from a file that were slow. This indicate system overload. Thresholds are controlled by read_backoff_* settings.', `ProfileEvent_ReadBackoff` UInt64 COMMENT 'Number of times the number of query processing threads was lowered due to slow reads.', `ProfileEvent_ReplicaPartialShutdown` UInt64 COMMENT 'How many times Replicated table has to deinitialize its state due to session expiration in ZooKeeper. The state is reinitialized every time when ZooKeeper is available again.', `ProfileEvent_SelectedParts` UInt64 COMMENT 'Number of data parts selected to read from a MergeTree table.', `ProfileEvent_SelectedPartsTotal` UInt64 COMMENT 'Number of total data parts before selecting which ones to read from a MergeTree table.', `ProfileEvent_SelectedRanges` UInt64 COMMENT 'Number of (non-adjacent) ranges in all data parts selected to read from a MergeTree table.', `ProfileEvent_SelectedMarks` UInt64 COMMENT 'Number of marks (index granules) selected to read from a MergeTree table.', `ProfileEvent_SelectedMarksTotal` UInt64 COMMENT 'Number of total marks (index granules) before selecting which ones to read from a MergeTree table.', `ProfileEvent_SelectedRows` UInt64 COMMENT 'Number of rows SELECTed from all tables.', `ProfileEvent_SelectedBytes` UInt64 COMMENT 'Number of bytes (uncompressed; for columns as they stored in memory) SELECTed from all tables.', `ProfileEvent_RowsReadByMainReader` UInt64 COMMENT 'Number of rows read from MergeTree tables by the main reader (after PREWHERE step).', `ProfileEvent_RowsReadByPrewhereReaders` UInt64 COMMENT 'Number of rows read from MergeTree tables (in total) by prewhere readers.', `ProfileEvent_LoadedDataParts` UInt64 COMMENT 'Number of data parts loaded by MergeTree tables during initialization.', `ProfileEvent_LoadedDataPartsMicroseconds` UInt64 COMMENT 'Microseconds spent by MergeTree tables for loading data parts during initialization.', `ProfileEvent_WaitMarksLoadMicroseconds` UInt64 COMMENT 'Time spent loading marks', `ProfileEvent_BackgroundLoadingMarksTasks` UInt64 COMMENT 'Number of background tasks for loading marks', `ProfileEvent_LoadingMarksTasksCanceled` UInt64 COMMENT 'Number of times background tasks for loading marks were canceled', `ProfileEvent_LoadedMarksFiles` UInt64 COMMENT 'Number of mark files loaded.', `ProfileEvent_LoadedMarksCount` UInt64 COMMENT 'Number of marks loaded (total across columns).', `ProfileEvent_LoadedMarksMemoryBytes` UInt64 COMMENT 'Size of in-memory representations of loaded marks.', `ProfileEvent_LoadedPrimaryIndexFiles` UInt64 COMMENT 'Number of primary index files loaded.', `ProfileEvent_LoadedPrimaryIndexRows` UInt64 COMMENT 'Number of rows of primary key loaded.', `ProfileEvent_LoadedPrimaryIndexBytes` UInt64 COMMENT 'Number of rows of primary key loaded.', `ProfileEvent_Merge` UInt64 COMMENT 'Number of launched background merges.', `ProfileEvent_MergeSourceParts` UInt64 COMMENT 'Number of source parts scheduled for merges.', `ProfileEvent_MergedRows` UInt64 COMMENT 'Rows read for background merges. This is the number of rows before merge.', `ProfileEvent_MergedColumns` UInt64 COMMENT 'Number of columns merged during the horizontal stage of merges.', `ProfileEvent_GatheredColumns` UInt64 COMMENT 'Number of columns gathered during the vertical stage of merges.', `ProfileEvent_MergedUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) that was read for background merges. This is the number before merge.', `ProfileEvent_MergeTotalMilliseconds` UInt64 COMMENT 'Total time spent for background merges', `ProfileEvent_MergeExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of background merges', `ProfileEvent_MergeHorizontalStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for horizontal stage of background merges', `ProfileEvent_MergeHorizontalStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of horizontal stage of background merges', `ProfileEvent_MergeVerticalStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for vertical stage of background merges', `ProfileEvent_MergeVerticalStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of vertical stage of background merges', `ProfileEvent_MergeProjectionStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for projection stage of background merges', `ProfileEvent_MergeProjectionStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of projection stage of background merges', `ProfileEvent_MergePrewarmStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for prewarm stage of background merges', `ProfileEvent_MergePrewarmStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of prewarm stage of background merges', `ProfileEvent_MergingSortedMilliseconds` UInt64 COMMENT 'Total time spent while merging sorted columns', `ProfileEvent_AggregatingSortedMilliseconds` UInt64 COMMENT 'Total time spent while aggregating sorted columns', `ProfileEvent_CollapsingSortedMilliseconds` UInt64 COMMENT 'Total time spent while collapsing sorted columns', `ProfileEvent_ReplacingSortedMilliseconds` UInt64 COMMENT 'Total time spent while replacing sorted columns', `ProfileEvent_SummingSortedMilliseconds` UInt64 COMMENT 'Total time spent while summing sorted columns', `ProfileEvent_VersionedCollapsingSortedMilliseconds` UInt64 COMMENT 'Total time spent while version collapsing sorted columns', `ProfileEvent_GatheringColumnMilliseconds` UInt64 COMMENT 'Total time spent while gathering columns for vertical merge', `ProfileEvent_MutationTotalParts` UInt64 COMMENT 'Number of total parts for which mutations tried to be applied', `ProfileEvent_MutationUntouchedParts` UInt64 COMMENT 'Number of total parts for which mutations tried to be applied but which was completely skipped according to predicate', `ProfileEvent_MutatedRows` UInt64 COMMENT 'Rows read for mutations. This is the number of rows before mutation', `ProfileEvent_MutatedUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) that was read for mutations. This is the number before mutation.', `ProfileEvent_MutationTotalMilliseconds` UInt64 COMMENT 'Total time spent for mutations.', `ProfileEvent_MutationExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of mutations.', `ProfileEvent_MutationAllPartColumns` UInt64 COMMENT 'Number of times when task to mutate all columns in part was created', `ProfileEvent_MutationSomePartColumns` UInt64 COMMENT 'Number of times when task to mutate some columns in part was created', `ProfileEvent_MutateTaskProjectionsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating projections in mutations', `ProfileEvent_MergeTreeDataWriterRows` UInt64 COMMENT 'Number of rows INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterCompressedBytes` UInt64 COMMENT 'Bytes written to filesystem for data INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterBlocks` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables. Each block forms a data part of level zero.', `ProfileEvent_MergeTreeDataWriterBlocksAlreadySorted` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables that appeared to be already sorted.', `ProfileEvent_MergeTreeDataWriterSkipIndicesCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating skip indices', `ProfileEvent_MergeTreeDataWriterStatisticsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating statistics', `ProfileEvent_MergeTreeDataWriterSortingBlocksMicroseconds` UInt64 COMMENT 'Time spent sorting blocks', `ProfileEvent_MergeTreeDataWriterMergingBlocksMicroseconds` UInt64 COMMENT 'Time spent merging input blocks (for special MergeTree engines)', `ProfileEvent_MergeTreeDataWriterProjectionsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating projections', `ProfileEvent_MergeTreeDataProjectionWriterSortingBlocksMicroseconds` UInt64 COMMENT 'Time spent sorting blocks (for projection it might be a key different from table\'s sorting key)', `ProfileEvent_MergeTreeDataProjectionWriterMergingBlocksMicroseconds` UInt64 COMMENT 'Time spent merging blocks', `ProfileEvent_InsertedWideParts` UInt64 COMMENT 'Number of parts inserted in Wide format.', `ProfileEvent_InsertedCompactParts` UInt64 COMMENT 'Number of parts inserted in Compact format.', `ProfileEvent_MergedIntoWideParts` UInt64 COMMENT 'Number of parts merged into Wide format.', `ProfileEvent_MergedIntoCompactParts` UInt64 COMMENT 'Number of parts merged into Compact format.', `ProfileEvent_MergeTreeDataProjectionWriterRows` UInt64 COMMENT 'Number of rows INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterCompressedBytes` UInt64 COMMENT 'Bytes written to filesystem for data INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterBlocks` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables projection. Each block forms a data part of level zero.', `ProfileEvent_MergeTreeDataProjectionWriterBlocksAlreadySorted` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables projection that appeared to be already sorted.', `ProfileEvent_CannotRemoveEphemeralNode` UInt64 COMMENT 'Number of times an error happened while trying to remove ephemeral node. This is not an issue, because our implementation of ZooKeeper library guarantee that the session will expire and the node will be removed.', `ProfileEvent_RegexpWithMultipleNeedlesCreated` UInt64 COMMENT 'Regular expressions with multiple needles (VectorScan library) compiled.', `ProfileEvent_RegexpWithMultipleNeedlesGlobalCacheHit` UInt64 COMMENT 'Number of times we fetched compiled regular expression with multiple needles (VectorScan library) from the global cache.', `ProfileEvent_RegexpWithMultipleNeedlesGlobalCacheMiss` UInt64 COMMENT 'Number of times we failed to fetch compiled regular expression with multiple needles (VectorScan library) from the global cache.', `ProfileEvent_RegexpLocalCacheHit` UInt64 COMMENT 'Number of times we fetched compiled regular expression from a local cache.', `ProfileEvent_RegexpLocalCacheMiss` UInt64 COMMENT 'Number of times we failed to fetch compiled regular expression from a local cache.', `ProfileEvent_ContextLock` UInt64 COMMENT 'Number of times the lock of Context was acquired or tried to acquire. This is global lock.', `ProfileEvent_ContextLockWaitMicroseconds` UInt64 COMMENT 'Context lock wait time in microseconds', `ProfileEvent_StorageBufferFlush` UInt64 COMMENT 'Number of times a buffer in a \'Buffer\' table was flushed.', `ProfileEvent_StorageBufferErrorOnFlush` UInt64 COMMENT 'Number of times a buffer in the \'Buffer\' table has not been able to flush due to error writing in the destination table.', `ProfileEvent_StorageBufferPassedAllMinThresholds` UInt64 COMMENT 'Number of times a criteria on min thresholds has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedTimeMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max time threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedRowsMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max rows threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedBytesMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max bytes threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedTimeFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on time has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferPassedRowsFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on rows has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferPassedBytesFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on bytes has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferLayerLockReadersWaitMilliseconds` UInt64 COMMENT 'Time for waiting for Buffer layer during reading.', `ProfileEvent_StorageBufferLayerLockWritersWaitMilliseconds` UInt64 COMMENT 'Time for waiting free Buffer layer to write to (can be used to tune Buffer layers).', `ProfileEvent_SystemLogErrorOnFlush` UInt64 COMMENT 'Number of times any of the system logs have failed to flush to the corresponding system table. Attempts to flush are repeated.', `ProfileEvent_DictCacheKeysRequested` UInt64 COMMENT 'Number of keys requested from the data source for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheKeysRequestedMiss` UInt64 COMMENT 'Number of keys requested from the data source for dictionaries of \'cache\' types but not found in the data source.', `ProfileEvent_DictCacheKeysRequestedFound` UInt64 COMMENT 'Number of keys requested from the data source for dictionaries of \'cache\' types and found in the data source.', `ProfileEvent_DictCacheKeysExpired` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and found in the cache but they were obsolete.', `ProfileEvent_DictCacheKeysNotFound` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and not found.', `ProfileEvent_DictCacheKeysHit` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and found in the cache.', `ProfileEvent_DictCacheRequestTimeNs` UInt64 COMMENT 'Number of nanoseconds spend in querying the external data sources for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheRequests` UInt64 COMMENT 'Number of bulk requests to the external data sources for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheLockWriteNs` UInt64 COMMENT 'Number of nanoseconds spend in waiting for write lock to update the data for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheLockReadNs` UInt64 COMMENT 'Number of nanoseconds spend in waiting for read lock to lookup the data for the dictionaries of \'cache\' types.', `ProfileEvent_DistributedSyncInsertionTimeoutExceeded` UInt64 COMMENT 'A timeout has exceeded while waiting for shards during synchronous insertion into a Distributed table (with \'distributed_foreground_insert\' = 1)', `ProfileEvent_DistributedAsyncInsertionFailures` UInt64 COMMENT 'Number of failures for asynchronous insertion into a Distributed table (with \'distributed_foreground_insert\' = 0)', `ProfileEvent_DataAfterMergeDiffersFromReplica` UInt64 COMMENT '\nNumber of times data after merge is not byte-identical to the data on another replicas. There could be several reasons:\n1. Using newer version of compression library after server update.\n2. Using another compression method.\n3. Non-deterministic compression algorithm (highly unlikely).\n4. Non-deterministic merge algorithm due to logical error in code.\n5. Data corruption in memory due to bug in code.\n6. Data corruption in memory due to hardware issue.\n7. Manual modification of source data after server startup.\n8. Manual modification of checksums stored in ZooKeeper.\n9. Part format related settings like \'enable_mixed_granularity_parts\' are different on different replicas.\nThe server successfully detected this situation and will download merged part from the replica to force the byte-identical result.\n', `ProfileEvent_DataAfterMutationDiffersFromReplica` UInt64 COMMENT 'Number of times data after mutation is not byte-identical to the data on other replicas. In addition to the reasons described in \'DataAfterMergeDiffersFromReplica\', it is also possible due to non-deterministic mutation.', `ProfileEvent_PolygonsAddedToPool` UInt64 COMMENT 'A polygon has been added to the cache (pool) for the \'pointInPolygon\' function.', `ProfileEvent_PolygonsInPoolAllocatedBytes` UInt64 COMMENT 'The number of bytes for polygons added to the cache (pool) for the \'pointInPolygon\' function.', `ProfileEvent_USearchAddCount` UInt64 COMMENT 'Number of vectors added to usearch indexes.', `ProfileEvent_USearchAddVisitedMembers` UInt64 COMMENT 'Number of nodes visited when adding vectors to usearch indexes.', `ProfileEvent_USearchAddComputedDistances` UInt64 COMMENT 'Number of times distance was computed when adding vectors to usearch indexes.', `ProfileEvent_USearchSearchCount` UInt64 COMMENT 'Number of search operations performed in usearch indexes.', `ProfileEvent_USearchSearchVisitedMembers` UInt64 COMMENT 'Number of nodes visited when searching in usearch indexes.', `ProfileEvent_USearchSearchComputedDistances` UInt64 COMMENT 'Number of times distance was computed when searching usearch indexes.', `ProfileEvent_RWLockAcquiredReadLocks` UInt64 COMMENT 'Number of times a read lock was acquired (in a heavy RWLock).', `ProfileEvent_RWLockAcquiredWriteLocks` UInt64 COMMENT 'Number of times a write lock was acquired (in a heavy RWLock).', `ProfileEvent_RWLockReadersWaitMilliseconds` UInt64 COMMENT 'Total time spent waiting for a read lock to be acquired (in a heavy RWLock).', `ProfileEvent_RWLockWritersWaitMilliseconds` UInt64 COMMENT 'Total time spent waiting for a write lock to be acquired (in a heavy RWLock).', `ProfileEvent_DNSError` UInt64 COMMENT 'Total count of errors in DNS resolution', `ProfileEvent_PartsLockHoldMicroseconds` UInt64 COMMENT 'Total time spent holding data parts lock in MergeTree tables', `ProfileEvent_PartsLockWaitMicroseconds` UInt64 COMMENT 'Total time spent waiting for data parts lock in MergeTree tables', `ProfileEvent_RealTimeMicroseconds` UInt64 COMMENT 'Total (wall clock) time spent in processing (queries and other tasks) threads (note that this is a sum).', `ProfileEvent_UserTimeMicroseconds` UInt64 COMMENT 'Total time spent in processing (queries and other tasks) threads executing CPU instructions in user mode. This includes time CPU pipeline was stalled due to main memory access, cache misses, branch mispredictions, hyper-threading, etc.', `ProfileEvent_SystemTimeMicroseconds` UInt64 COMMENT 'Total time spent in processing (queries and other tasks) threads executing CPU instructions in OS kernel mode. This is time spent in syscalls, excluding waiting time during blocking syscalls.', `ProfileEvent_MemoryOvercommitWaitTimeMicroseconds` UInt64 COMMENT 'Total time spent in waiting for memory to be freed in OvercommitTracker.', `ProfileEvent_MemoryAllocatorPurge` UInt64 COMMENT 'Total number of times memory allocator purge was requested', `ProfileEvent_MemoryAllocatorPurgeTimeMicroseconds` UInt64 COMMENT 'Total number of times memory allocator purge was requested', `ProfileEvent_SoftPageFaults` UInt64 COMMENT 'The number of soft page faults in query execution threads. Soft page fault usually means a miss in the memory allocator cache, which requires a new memory mapping from the OS and subsequent allocation of a page of physical memory.', `ProfileEvent_HardPageFaults` UInt64 COMMENT 'The number of hard page faults in query execution threads. High values indicate either that you forgot to turn off swap on your server, or eviction of memory pages of the ClickHouse binary during very high memory pressure, or successful usage of the \'mmap\' read method for the tables data.', `ProfileEvent_OSIOWaitMicroseconds` UInt64 COMMENT 'Total time a thread spent waiting for a result of IO operation, from the OS point of view. This is real IO that doesn\'t include page cache.', `ProfileEvent_OSCPUWaitMicroseconds` UInt64 COMMENT 'Total time a thread was ready for execution but waiting to be scheduled by OS, from the OS point of view.', `ProfileEvent_OSCPUVirtualTimeMicroseconds` UInt64 COMMENT 'CPU time spent seen by OS. Does not include involuntary waits due to virtualization.', `ProfileEvent_OSReadBytes` UInt64 COMMENT 'Number of bytes read from disks or block devices. Doesn\'t include bytes read from page cache. May include excessive data due to block size, readahead, etc.', `ProfileEvent_OSWriteBytes` UInt64 COMMENT 'Number of bytes written to disks or block devices. Doesn\'t include bytes that are in page cache dirty pages. May not include data that was written by OS asynchronously.', `ProfileEvent_OSReadChars` UInt64 COMMENT 'Number of bytes read from filesystem, including page cache.', `ProfileEvent_OSWriteChars` UInt64 COMMENT 'Number of bytes written to filesystem, including page cache.', `ProfileEvent_ParallelReplicasHandleRequestMicroseconds` UInt64 COMMENT 'Time spent processing requests for marks from replicas', `ProfileEvent_ParallelReplicasHandleAnnouncementMicroseconds` UInt64 COMMENT 'Time spent processing replicas announcements', `ProfileEvent_ParallelReplicasAnnouncementMicroseconds` UInt64 COMMENT 'Time spent to send an announcement', `ProfileEvent_ParallelReplicasReadRequestMicroseconds` UInt64 COMMENT 'Time spent for read requests', `ProfileEvent_ParallelReplicasReadAssignedMarks` UInt64 COMMENT 'Sum across all replicas of how many of scheduled marks were assigned by consistent hash', `ProfileEvent_ParallelReplicasReadUnassignedMarks` UInt64 COMMENT 'Sum across all replicas of how many unassigned marks were scheduled', `ProfileEvent_ParallelReplicasReadAssignedForStealingMarks` UInt64 COMMENT 'Sum across all replicas of how many of scheduled marks were assigned for stealing by consistent hash', `ProfileEvent_ParallelReplicasReadMarks` UInt64 COMMENT 'How many marks were read by the given replica', `ProfileEvent_ParallelReplicasStealingByHashMicroseconds` UInt64 COMMENT 'Time spent collecting segments meant for stealing by hash', `ProfileEvent_ParallelReplicasProcessingPartsMicroseconds` UInt64 COMMENT 'Time spent processing data parts', `ProfileEvent_ParallelReplicasStealingLeftoversMicroseconds` UInt64 COMMENT 'Time spent collecting orphaned segments', `ProfileEvent_ParallelReplicasCollectingOwnedSegmentsMicroseconds` UInt64 COMMENT 'Time spent collecting segments meant by hash', `ProfileEvent_ParallelReplicasNumRequests` UInt64 COMMENT 'Number of requests to the initiator.', `ProfileEvent_ParallelReplicasDeniedRequests` UInt64 COMMENT 'Number of completely denied requests to the initiator', `ProfileEvent_CacheWarmerBytesDownloaded` UInt64 COMMENT 'Amount of data fetched into filesystem cache by dedicated background threads.', `ProfileEvent_CacheWarmerDataPartsDownloaded` UInt64 COMMENT 'Number of data parts that were fully fetched by CacheWarmer.', `ProfileEvent_IgnoredColdParts` UInt64 COMMENT 'See setting ignore_cold_parts_seconds. Number of times read queries ignored very new parts that weren\'t pulled into cache by CacheWarmer yet.', `ProfileEvent_PreferredWarmedUnmergedParts` UInt64 COMMENT 'See setting prefer_warmed_unmerged_parts_seconds. Number of times read queries used outdated pre-merge parts that are in cache instead of merged part that wasn\'t pulled into cache by CacheWarmer yet.', `ProfileEvent_PerfCPUCycles` UInt64 COMMENT 'Total cycles. Be wary of what happens during CPU frequency scaling.', `ProfileEvent_PerfInstructions` UInt64 COMMENT 'Retired instructions. Be careful, these can be affected by various issues, most notably hardware interrupt counts.', `ProfileEvent_PerfCacheReferences` UInt64 COMMENT 'Cache accesses. Usually, this indicates Last Level Cache accesses, but this may vary depending on your CPU. This may include prefetches and coherency messages; again this depends on the design of your CPU.', `ProfileEvent_PerfCacheMisses` UInt64 COMMENT 'Cache misses. Usually this indicates Last Level Cache misses; this is intended to be used in conjunction with the PERFCOUNTHWCACHEREFERENCES event to calculate cache miss rates.', `ProfileEvent_PerfBranchInstructions` UInt64 COMMENT 'Retired branch instructions. Prior to Linux 2.6.35, this used the wrong event on AMD processors.', `ProfileEvent_PerfBranchMisses` UInt64 COMMENT 'Mispredicted branch instructions.', `ProfileEvent_PerfBusCycles` UInt64 COMMENT 'Bus cycles, which can be different from total cycles.', `ProfileEvent_PerfStalledCyclesFrontend` UInt64 COMMENT 'Stalled cycles during issue.', `ProfileEvent_PerfStalledCyclesBackend` UInt64 COMMENT 'Stalled cycles during retirement.', `ProfileEvent_PerfRefCPUCycles` UInt64 COMMENT 'Total cycles; not affected by CPU frequency scaling.', `ProfileEvent_PerfCPUClock` UInt64 COMMENT 'The CPU clock, a high-resolution per-CPU timer', `ProfileEvent_PerfTaskClock` UInt64 COMMENT 'A clock count specific to the task that is running', `ProfileEvent_PerfContextSwitches` UInt64 COMMENT 'Number of context switches', `ProfileEvent_PerfCPUMigrations` UInt64 COMMENT 'Number of times the process has migrated to a new CPU', `ProfileEvent_PerfAlignmentFaults` UInt64 COMMENT 'Number of alignment faults. These happen when unaligned memory accesses happen; the kernel can handle these but it reduces performance. This happens only on some architectures (never on x86).', `ProfileEvent_PerfEmulationFaults` UInt64 COMMENT 'Number of emulation faults. The kernel sometimes traps on unimplemented instructions and emulates them for user space. This can negatively impact performance.', `ProfileEvent_PerfMinEnabledTime` UInt64 COMMENT 'For all events, minimum time that an event was enabled. Used to track event multiplexing influence', `ProfileEvent_PerfMinEnabledRunningTime` UInt64 COMMENT 'Running time for event with minimum enabled time. Used to track the amount of event multiplexing', `ProfileEvent_PerfDataTLBReferences` UInt64 COMMENT 'Data TLB references', `ProfileEvent_PerfDataTLBMisses` UInt64 COMMENT 'Data TLB misses', `ProfileEvent_PerfInstructionTLBReferences` UInt64 COMMENT 'Instruction TLB references', `ProfileEvent_PerfInstructionTLBMisses` UInt64 COMMENT 'Instruction TLB misses', `ProfileEvent_PerfLocalMemoryReferences` UInt64 COMMENT 'Local NUMA node memory reads', `ProfileEvent_PerfLocalMemoryMisses` UInt64 COMMENT 'Local NUMA node memory read misses', `ProfileEvent_CannotWriteToWriteBufferDiscard` UInt64 COMMENT 'Number of stack traces dropped by query profiler or signal handler because pipe is full or cannot write to pipe.', `ProfileEvent_QueryProfilerSignalOverruns` UInt64 COMMENT 'Number of times we drop processing of a query profiler signal due to overrun plus the number of signals that OS has not delivered due to overrun.', `ProfileEvent_QueryProfilerConcurrencyOverruns` UInt64 COMMENT 'Number of times we drop processing of a query profiler signal due to too many concurrent query profilers in other threads, which may indicate overload.', `ProfileEvent_QueryProfilerRuns` UInt64 COMMENT 'Number of times QueryProfiler had been run.', `ProfileEvent_QueryProfilerErrors` UInt64 COMMENT 'Invalid memory accesses during asynchronous stack unwinding.', `ProfileEvent_CreatedLogEntryForMerge` UInt64 COMMENT 'Successfully created log entry to merge parts in ReplicatedMergeTree.', `ProfileEvent_NotCreatedLogEntryForMerge` UInt64 COMMENT 'Log entry to merge parts in ReplicatedMergeTree is not created due to concurrent log update by another replica.', `ProfileEvent_CreatedLogEntryForMutation` UInt64 COMMENT 'Successfully created log entry to mutate parts in ReplicatedMergeTree.', `ProfileEvent_NotCreatedLogEntryForMutation` UInt64 COMMENT 'Log entry to mutate parts in ReplicatedMergeTree is not created due to concurrent log update by another replica.', `ProfileEvent_S3ReadMicroseconds` UInt64 COMMENT 'Time of GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsCount` UInt64 COMMENT 'Number of GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsRedirects` UInt64 COMMENT 'Number of redirects in GET and HEAD requests to S3 storage.', `ProfileEvent_S3WriteMicroseconds` UInt64 COMMENT 'Time of POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsCount` UInt64 COMMENT 'Number of POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsRedirects` UInt64 COMMENT 'Number of redirects in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_DiskS3ReadMicroseconds` UInt64 COMMENT 'Time of GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsCount` UInt64 COMMENT 'Number of GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsRedirects` UInt64 COMMENT 'Number of redirects in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteMicroseconds` UInt64 COMMENT 'Time of POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsCount` UInt64 COMMENT 'Number of POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsRedirects` UInt64 COMMENT 'Number of redirects in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_S3DeleteObjects` UInt64 COMMENT 'Number of S3 API DeleteObject(s) calls.', `ProfileEvent_S3CopyObject` UInt64 COMMENT 'Number of S3 API CopyObject calls.', `ProfileEvent_S3ListObjects` UInt64 COMMENT 'Number of S3 API ListObjects calls.', `ProfileEvent_S3HeadObject` UInt64 COMMENT 'Number of S3 API HeadObject calls.', `ProfileEvent_S3GetObjectAttributes` UInt64 COMMENT 'Number of S3 API GetObjectAttributes calls.', `ProfileEvent_S3CreateMultipartUpload` UInt64 COMMENT 'Number of S3 API CreateMultipartUpload calls.', `ProfileEvent_S3UploadPartCopy` UInt64 COMMENT 'Number of S3 API UploadPartCopy calls.', `ProfileEvent_S3UploadPart` UInt64 COMMENT 'Number of S3 API UploadPart calls.', `ProfileEvent_S3AbortMultipartUpload` UInt64 COMMENT 'Number of S3 API AbortMultipartUpload calls.', `ProfileEvent_S3CompleteMultipartUpload` UInt64 COMMENT 'Number of S3 API CompleteMultipartUpload calls.', `ProfileEvent_S3PutObject` UInt64 COMMENT 'Number of S3 API PutObject calls.', `ProfileEvent_S3GetObject` UInt64 COMMENT 'Number of S3 API GetObject calls.', `ProfileEvent_DiskS3DeleteObjects` UInt64 COMMENT 'Number of DiskS3 API DeleteObject(s) calls.', `ProfileEvent_DiskS3CopyObject` UInt64 COMMENT 'Number of DiskS3 API CopyObject calls.', `ProfileEvent_DiskS3ListObjects` UInt64 COMMENT 'Number of DiskS3 API ListObjects calls.', `ProfileEvent_DiskS3HeadObject` UInt64 COMMENT 'Number of DiskS3 API HeadObject calls.', `ProfileEvent_DiskS3GetObjectAttributes` UInt64 COMMENT 'Number of DiskS3 API GetObjectAttributes calls.', `ProfileEvent_DiskS3CreateMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API CreateMultipartUpload calls.', `ProfileEvent_DiskS3UploadPartCopy` UInt64 COMMENT 'Number of DiskS3 API UploadPartCopy calls.', `ProfileEvent_DiskS3UploadPart` UInt64 COMMENT 'Number of DiskS3 API UploadPart calls.', `ProfileEvent_DiskS3AbortMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API AbortMultipartUpload calls.', `ProfileEvent_DiskS3CompleteMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API CompleteMultipartUpload calls.', `ProfileEvent_DiskS3PutObject` UInt64 COMMENT 'Number of DiskS3 API PutObject calls.', `ProfileEvent_DiskS3GetObject` UInt64 COMMENT 'Number of DiskS3 API GetObject calls.', `ProfileEvent_DiskPlainRewritableAzureDirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for AzureObjectStorage.', `ProfileEvent_DiskPlainRewritableAzureDirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for AzureObjectStorage.', `ProfileEvent_DiskPlainRewritableLocalDirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for LocalObjectStorage.', `ProfileEvent_DiskPlainRewritableLocalDirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for LocalObjectStorage.', `ProfileEvent_DiskPlainRewritableS3DirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for S3ObjectStorage.', `ProfileEvent_DiskPlainRewritableS3DirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for S3ObjectStorage.', `ProfileEvent_S3Clients` UInt64 COMMENT 'Number of created S3 clients.', `ProfileEvent_TinyS3Clients` UInt64 COMMENT 'Number of S3 clients copies which reuse an existing auth provider from another client.', `ProfileEvent_EngineFileLikeReadFiles` UInt64 COMMENT 'Number of files read in table engines working with files (like File/S3/URL/HDFS).', `ProfileEvent_ReadBufferFromS3Microseconds` UInt64 COMMENT 'Time spent on reading from S3.', `ProfileEvent_ReadBufferFromS3InitMicroseconds` UInt64 COMMENT 'Time spent initializing connection to S3.', `ProfileEvent_ReadBufferFromS3Bytes` UInt64 COMMENT 'Bytes read from S3.', `ProfileEvent_ReadBufferFromS3RequestsErrors` UInt64 COMMENT 'Number of exceptions while reading from S3.', `ProfileEvent_WriteBufferFromS3Microseconds` UInt64 COMMENT 'Time spent on writing to S3.', `ProfileEvent_WriteBufferFromS3Bytes` UInt64 COMMENT 'Bytes written to S3.', `ProfileEvent_WriteBufferFromS3RequestsErrors` UInt64 COMMENT 'Number of exceptions while writing to S3.', `ProfileEvent_WriteBufferFromS3WaitInflightLimitMicroseconds` UInt64 COMMENT 'Time spent on waiting while some of the current requests are done when its number reached the limit defined by s3_max_inflight_parts_for_one_file.', `ProfileEvent_QueryMemoryLimitExceeded` UInt64 COMMENT 'Number of times when memory limit exceeded for query.', `ProfileEvent_AzureGetObject` UInt64 COMMENT 'Number of Azure API GetObject calls.', `ProfileEvent_AzureUpload` UInt64 COMMENT 'Number of Azure blob storage API Upload calls', `ProfileEvent_AzureStageBlock` UInt64 COMMENT 'Number of Azure blob storage API StageBlock calls', `ProfileEvent_AzureCommitBlockList` UInt64 COMMENT 'Number of Azure blob storage API CommitBlockList calls', `ProfileEvent_AzureCopyObject` UInt64 COMMENT 'Number of Azure blob storage API CopyObject calls', `ProfileEvent_AzureDeleteObjects` UInt64 COMMENT 'Number of Azure blob storage API DeleteObject(s) calls.', `ProfileEvent_AzureListObjects` UInt64 COMMENT 'Number of Azure blob storage API ListObjects calls.', `ProfileEvent_AzureGetProperties` UInt64 COMMENT 'Number of Azure blob storage API GetProperties calls.', `ProfileEvent_AzureCreateContainer` UInt64 COMMENT 'Number of Azure blob storage API CreateContainer calls.', `ProfileEvent_DiskAzureGetObject` UInt64 COMMENT 'Number of Disk Azure API GetObject calls.', `ProfileEvent_DiskAzureUpload` UInt64 COMMENT 'Number of Disk Azure blob storage API Upload calls', `ProfileEvent_DiskAzureStageBlock` UInt64 COMMENT 'Number of Disk Azure blob storage API StageBlock calls', `ProfileEvent_DiskAzureCommitBlockList` UInt64 COMMENT 'Number of Disk Azure blob storage API CommitBlockList calls', `ProfileEvent_DiskAzureCopyObject` UInt64 COMMENT 'Number of Disk Azure blob storage API CopyObject calls', `ProfileEvent_DiskAzureListObjects` UInt64 COMMENT 'Number of Disk Azure blob storage API ListObjects calls.', `ProfileEvent_DiskAzureDeleteObjects` UInt64 COMMENT 'Number of Azure blob storage API DeleteObject(s) calls.', `ProfileEvent_DiskAzureGetProperties` UInt64 COMMENT 'Number of Disk Azure blob storage API GetProperties calls.', `ProfileEvent_DiskAzureCreateContainer` UInt64 COMMENT 'Number of Disk Azure blob storage API CreateContainer calls.', `ProfileEvent_ReadBufferFromAzureMicroseconds` UInt64 COMMENT 'Time spent on reading from Azure.', `ProfileEvent_ReadBufferFromAzureInitMicroseconds` UInt64 COMMENT 'Time spent initializing connection to Azure.', `ProfileEvent_ReadBufferFromAzureBytes` UInt64 COMMENT 'Bytes read from Azure.', `ProfileEvent_ReadBufferFromAzureRequestsErrors` UInt64 COMMENT 'Number of exceptions while reading from Azure', `ProfileEvent_CachedReadBufferReadFromCacheHits` UInt64 COMMENT 'Number of times the read from filesystem cache hit the cache.', `ProfileEvent_CachedReadBufferReadFromCacheMisses` UInt64 COMMENT 'Number of times the read from filesystem cache miss the cache.', `ProfileEvent_CachedReadBufferReadFromSourceMicroseconds` UInt64 COMMENT 'Time reading from filesystem cache source (from remote filesystem, etc)', `ProfileEvent_CachedReadBufferReadFromCacheMicroseconds` UInt64 COMMENT 'Time reading from filesystem cache', `ProfileEvent_CachedReadBufferReadFromSourceBytes` UInt64 COMMENT 'Bytes read from filesystem cache source (from remote fs, etc)', `ProfileEvent_CachedReadBufferReadFromCacheBytes` UInt64 COMMENT 'Bytes read from filesystem cache', `ProfileEvent_CachedReadBufferPredownloadedBytes` UInt64 COMMENT 'Bytes read from filesystem cache source. Cache segments are read from left to right as a whole, it might be that we need to predownload some part of the segment irrelevant for the current task just to get to the needed data', `ProfileEvent_CachedReadBufferCacheWriteBytes` UInt64 COMMENT 'Bytes written from source (remote fs, etc) to filesystem cache', `ProfileEvent_CachedReadBufferCacheWriteMicroseconds` UInt64 COMMENT 'Time spent writing data into filesystem cache', `ProfileEvent_CachedReadBufferCreateBufferMicroseconds` UInt64 COMMENT 'Prepare buffer time', `ProfileEvent_CachedWriteBufferCacheWriteBytes` UInt64 COMMENT 'Bytes written from source (remote fs, etc) to filesystem cache', `ProfileEvent_CachedWriteBufferCacheWriteMicroseconds` UInt64 COMMENT 'Time spent writing data into filesystem cache', `ProfileEvent_FilesystemCacheLoadMetadataMicroseconds` UInt64 COMMENT 'Time spent loading filesystem cache metadata', `ProfileEvent_FilesystemCacheEvictedBytes` UInt64 COMMENT 'Number of bytes evicted from filesystem cache', `ProfileEvent_FilesystemCacheEvictedFileSegments` UInt64 COMMENT 'Number of file segments evicted from filesystem cache', `ProfileEvent_FilesystemCacheBackgroundDownloadQueuePush` UInt64 COMMENT 'Number of file segments sent for background download in filesystem cache', `ProfileEvent_FilesystemCacheEvictionSkippedFileSegments` UInt64 COMMENT 'Number of file segments skipped for eviction because of being in unreleasable state', `ProfileEvent_FilesystemCacheEvictionSkippedEvictingFileSegments` UInt64 COMMENT 'Number of file segments skipped for eviction because of being in evicting state', `ProfileEvent_FilesystemCacheEvictionTries` UInt64 COMMENT 'Number of filesystem cache eviction attempts', `ProfileEvent_FilesystemCacheLockKeyMicroseconds` UInt64 COMMENT 'Lock cache key time', `ProfileEvent_FilesystemCacheLockMetadataMicroseconds` UInt64 COMMENT 'Lock filesystem cache metadata time', `ProfileEvent_FilesystemCacheLockCacheMicroseconds` UInt64 COMMENT 'Lock filesystem cache time', `ProfileEvent_FilesystemCacheReserveMicroseconds` UInt64 COMMENT 'Filesystem cache space reservation time', `ProfileEvent_FilesystemCacheEvictMicroseconds` UInt64 COMMENT 'Filesystem cache eviction time', `ProfileEvent_FilesystemCacheGetOrSetMicroseconds` UInt64 COMMENT 'Filesystem cache getOrSet() time', `ProfileEvent_FilesystemCacheGetMicroseconds` UInt64 COMMENT 'Filesystem cache get() time', `ProfileEvent_FileSegmentWaitMicroseconds` UInt64 COMMENT 'Wait on DOWNLOADING state', `ProfileEvent_FileSegmentCompleteMicroseconds` UInt64 COMMENT 'Duration of FileSegment::complete() in filesystem cache', `ProfileEvent_FileSegmentLockMicroseconds` UInt64 COMMENT 'Lock file segment time', `ProfileEvent_FileSegmentWriteMicroseconds` UInt64 COMMENT 'File segment write() time', `ProfileEvent_FileSegmentUseMicroseconds` UInt64 COMMENT 'File segment use() time', `ProfileEvent_FileSegmentRemoveMicroseconds` UInt64 COMMENT 'File segment remove() time', `ProfileEvent_FileSegmentHolderCompleteMicroseconds` UInt64 COMMENT 'File segments holder complete() time', `ProfileEvent_FileSegmentFailToIncreasePriority` UInt64 COMMENT 'Number of times the priority was not increased due to a high contention on the cache lock', `ProfileEvent_FilesystemCacheFailToReserveSpaceBecauseOfLockContention` UInt64 COMMENT 'Number of times space reservation was skipped due to a high contention on the cache lock', `ProfileEvent_FilesystemCacheFailToReserveSpaceBecauseOfCacheResize` UInt64 COMMENT 'Number of times space reservation was skipped due to the cache is being resized', `ProfileEvent_FilesystemCacheHoldFileSegments` UInt64 COMMENT 'Filesystem cache file segments count, which were hold', `ProfileEvent_FilesystemCacheUnusedHoldFileSegments` UInt64 COMMENT 'Filesystem cache file segments count, which were hold, but not used (because of seek or LIMIT n, etc)', `ProfileEvent_FilesystemCacheFreeSpaceKeepingThreadRun` UInt64 COMMENT 'Number of times background thread executed free space keeping job', `ProfileEvent_FilesystemCacheFreeSpaceKeepingThreadWorkMilliseconds` UInt64 COMMENT 'Time for which background thread executed free space keeping job', `ProfileEvent_RemoteFSSeeks` UInt64 COMMENT 'Total number of seeks for async buffer', `ProfileEvent_RemoteFSPrefetches` UInt64 COMMENT 'Number of prefetches made with asynchronous reading from remote filesystem', `ProfileEvent_RemoteFSCancelledPrefetches` UInt64 COMMENT 'Number of cancelled prefecthes (because of seek)', `ProfileEvent_RemoteFSUnusedPrefetches` UInt64 COMMENT 'Number of prefetches pending at buffer destruction', `ProfileEvent_RemoteFSPrefetchedReads` UInt64 COMMENT 'Number of reads from prefecthed buffer', `ProfileEvent_RemoteFSPrefetchedBytes` UInt64 COMMENT 'Number of bytes from prefecthed buffer', `ProfileEvent_RemoteFSUnprefetchedReads` UInt64 COMMENT 'Number of reads from unprefetched buffer', `ProfileEvent_RemoteFSUnprefetchedBytes` UInt64 COMMENT 'Number of bytes from unprefetched buffer', `ProfileEvent_RemoteFSLazySeeks` UInt64 COMMENT 'Number of lazy seeks', `ProfileEvent_RemoteFSSeeksWithReset` UInt64 COMMENT 'Number of seeks which lead to a new connection', `ProfileEvent_RemoteFSBuffers` UInt64 COMMENT 'Number of buffers created for asynchronous reading from remote filesystem', `ProfileEvent_MergeTreePrefetchedReadPoolInit` UInt64 COMMENT 'Time spent preparing tasks in MergeTreePrefetchedReadPool', `ProfileEvent_WaitPrefetchTaskMicroseconds` UInt64 COMMENT 'Time spend waiting for prefetched reader', `ProfileEvent_ThreadpoolReaderTaskMicroseconds` UInt64 COMMENT 'Time spent getting the data in asynchronous reading', `ProfileEvent_ThreadpoolReaderPrepareMicroseconds` UInt64 COMMENT 'Time spent on preparation (e.g. call to reader seek() method)', `ProfileEvent_ThreadpoolReaderReadBytes` UInt64 COMMENT 'Bytes read from a threadpool task in asynchronous reading', `ProfileEvent_ThreadpoolReaderSubmit` UInt64 COMMENT 'Bytes read from a threadpool task in asynchronous reading', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronously` UInt64 COMMENT 'How many times we haven\'t scheduled a task on the thread pool and read synchronously instead', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronouslyBytes` UInt64 COMMENT 'How many bytes were read synchronously', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronouslyMicroseconds` UInt64 COMMENT 'How much time we spent reading synchronously', `ProfileEvent_ThreadpoolReaderSubmitLookupInCacheMicroseconds` UInt64 COMMENT 'How much time we spent checking if content is cached', `ProfileEvent_AsynchronousReaderIgnoredBytes` UInt64 COMMENT 'Number of bytes ignored during asynchronous reading', `ProfileEvent_FileSegmentWaitReadBufferMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend waiting for internal read buffer (includes cache waiting)', `ProfileEvent_FileSegmentReadMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend reading from file', `ProfileEvent_FileSegmentCacheWriteMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend writing data to cache', `ProfileEvent_FileSegmentPredownloadMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spent pre-downloading data to cache (pre-downloading - finishing file segment download (after someone who failed to do that) up to the point current thread was requested to do)', `ProfileEvent_FileSegmentUsedBytes` UInt64 COMMENT 'Metric per file segment. How many bytes were actually used from current file segment', `ProfileEvent_ReadBufferSeekCancelConnection` UInt64 COMMENT 'Number of seeks which lead to new connection (s3, http)', `ProfileEvent_SleepFunctionCalls` UInt64 COMMENT 'Number of times a sleep function (sleep, sleepEachRow) has been called.', `ProfileEvent_SleepFunctionMicroseconds` UInt64 COMMENT 'Time set to sleep in a sleep function (sleep, sleepEachRow).', `ProfileEvent_SleepFunctionElapsedMicroseconds` UInt64 COMMENT 'Time spent sleeping in a sleep function (sleep, sleepEachRow).', `ProfileEvent_ThreadPoolReaderPageCacheHit` UInt64 COMMENT 'Number of times the read inside ThreadPoolReader was done from the page cache.', `ProfileEvent_ThreadPoolReaderPageCacheHitBytes` UInt64 COMMENT 'Number of bytes read inside ThreadPoolReader when it was done from the page cache.', `ProfileEvent_ThreadPoolReaderPageCacheHitElapsedMicroseconds` UInt64 COMMENT 'Time spent reading data from page cache in ThreadPoolReader.', `ProfileEvent_ThreadPoolReaderPageCacheMiss` UInt64 COMMENT 'Number of times the read inside ThreadPoolReader was not done from page cache and was hand off to thread pool.', `ProfileEvent_ThreadPoolReaderPageCacheMissBytes` UInt64 COMMENT 'Number of bytes read inside ThreadPoolReader when read was not done from page cache and was hand off to thread pool.', `ProfileEvent_ThreadPoolReaderPageCacheMissElapsedMicroseconds` UInt64 COMMENT 'Time spent reading data inside the asynchronous job in ThreadPoolReader - when read was not done from the page cache.', `ProfileEvent_AsynchronousReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for asynchronous reads in asynchronous local read.', `ProfileEvent_SynchronousReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for synchronous reads in asynchronous local read.', `ProfileEvent_AsynchronousRemoteReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for asynchronous remote reads.', `ProfileEvent_SynchronousRemoteReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for synchronous remote reads.', `ProfileEvent_ExternalDataSourceLocalCacheReadBytes` UInt64 COMMENT 'Bytes read from local cache buffer in RemoteReadBufferCache', `ProfileEvent_MainConfigLoads` UInt64 COMMENT 'Number of times the main configuration was reloaded.', `ProfileEvent_AggregationPreallocatedElementsInHashTables` UInt64 COMMENT 'How many elements were preallocated in hash tables for aggregation.', `ProfileEvent_AggregationHashTablesInitializedAsTwoLevel` UInt64 COMMENT 'How many hash tables were inited as two-level for aggregation.', `ProfileEvent_AggregationOptimizedEqualRangesOfKeys` UInt64 COMMENT 'For how many blocks optimization of equal ranges of keys was applied', `ProfileEvent_HashJoinPreallocatedElementsInHashTables` UInt64 COMMENT 'How many elements were preallocated in hash tables for hash join.', `ProfileEvent_MetadataFromKeeperCacheHit` UInt64 COMMENT 'Number of times an object storage metadata request was answered from cache without making request to Keeper', `ProfileEvent_MetadataFromKeeperCacheMiss` UInt64 COMMENT 'Number of times an object storage metadata request had to be answered from Keeper', `ProfileEvent_MetadataFromKeeperCacheUpdateMicroseconds` UInt64 COMMENT 'Total time spent in updating the cache including waiting for responses from Keeper', `ProfileEvent_MetadataFromKeeperUpdateCacheOneLevel` UInt64 COMMENT 'Number of times a cache update for one level of directory tree was done', `ProfileEvent_MetadataFromKeeperTransactionCommit` UInt64 COMMENT 'Number of times metadata transaction commit was attempted', `ProfileEvent_MetadataFromKeeperTransactionCommitRetry` UInt64 COMMENT 'Number of times metadata transaction commit was retried', `ProfileEvent_MetadataFromKeeperCleanupTransactionCommit` UInt64 COMMENT 'Number of times metadata transaction commit for deleted objects cleanup was attempted', `ProfileEvent_MetadataFromKeeperCleanupTransactionCommitRetry` UInt64 COMMENT 'Number of times metadata transaction commit for deleted objects cleanup was retried', `ProfileEvent_MetadataFromKeeperOperations` UInt64 COMMENT 'Number of times a request was made to Keeper', `ProfileEvent_MetadataFromKeeperIndividualOperations` UInt64 COMMENT 'Number of paths read or written by single or multi requests to Keeper', `ProfileEvent_MetadataFromKeeperReconnects` UInt64 COMMENT 'Number of times a reconnect to Keeper was done', `ProfileEvent_MetadataFromKeeperBackgroundCleanupObjects` UInt64 COMMENT 'Number of times a old deleted object clean up was performed by background task', `ProfileEvent_MetadataFromKeeperBackgroundCleanupTransactions` UInt64 COMMENT 'Number of times old transaction idempotency token was cleaned up by background task', `ProfileEvent_MetadataFromKeeperBackgroundCleanupErrors` UInt64 COMMENT 'Number of times an error was encountered in background cleanup task', `ProfileEvent_SharedMergeTreeMetadataCacheHintLoadedFromCache` UInt64 COMMENT 'Number of times metadata cache hint was found without going to Keeper', `ProfileEvent_KafkaRebalanceRevocations` UInt64 COMMENT 'Number of partition revocations (the first stage of consumer group rebalance)', `ProfileEvent_KafkaRebalanceAssignments` UInt64 COMMENT 'Number of partition assignments (the final stage of consumer group rebalance)', `ProfileEvent_KafkaRebalanceErrors` UInt64 COMMENT 'Number of failed consumer group rebalances', `ProfileEvent_KafkaMessagesPolled` UInt64 COMMENT 'Number of Kafka messages polled from librdkafka to ClickHouse', `ProfileEvent_KafkaMessagesRead` UInt64 COMMENT 'Number of Kafka messages already processed by ClickHouse', `ProfileEvent_KafkaMessagesFailed` UInt64 COMMENT 'Number of Kafka messages ClickHouse failed to parse', `ProfileEvent_KafkaRowsRead` UInt64 COMMENT 'Number of rows parsed from Kafka messages', `ProfileEvent_KafkaRowsRejected` UInt64 COMMENT 'Number of parsed rows which were later rejected (due to rebalances / errors or similar reasons). Those rows will be consumed again after the rebalance.', `ProfileEvent_KafkaDirectReads` UInt64 COMMENT 'Number of direct selects from Kafka tables since server start', `ProfileEvent_KafkaBackgroundReads` UInt64 COMMENT 'Number of background reads populating materialized views from Kafka since server start', `ProfileEvent_KafkaCommits` UInt64 COMMENT 'Number of successful commits of consumed offsets to Kafka (normally should be the same as KafkaBackgroundReads)', `ProfileEvent_KafkaCommitFailures` UInt64 COMMENT 'Number of failed commits of consumed offsets to Kafka (usually is a sign of some data duplication)', `ProfileEvent_KafkaConsumerErrors` UInt64 COMMENT 'Number of errors reported by librdkafka during polls', `ProfileEvent_KafkaWrites` UInt64 COMMENT 'Number of writes (inserts) to Kafka tables ', `ProfileEvent_KafkaRowsWritten` UInt64 COMMENT 'Number of rows inserted into Kafka tables', `ProfileEvent_KafkaProducerFlushes` UInt64 COMMENT 'Number of explicit flushes to Kafka producer', `ProfileEvent_KafkaMessagesProduced` UInt64 COMMENT 'Number of messages produced to Kafka', `ProfileEvent_KafkaProducerErrors` UInt64 COMMENT 'Number of errors during producing the messages to Kafka', `ProfileEvent_ScalarSubqueriesGlobalCacheHit` UInt64 COMMENT 'Number of times a read from a scalar subquery was done using the global cache', `ProfileEvent_ScalarSubqueriesLocalCacheHit` UInt64 COMMENT 'Number of times a read from a scalar subquery was done using the local cache', `ProfileEvent_ScalarSubqueriesCacheMiss` UInt64 COMMENT 'Number of times a read from a scalar subquery was not cached and had to be calculated completely', `ProfileEvent_SchemaInferenceCacheHits` UInt64 COMMENT 'Number of times the requested source is found in schema cache', `ProfileEvent_SchemaInferenceCacheSchemaHits` UInt64 COMMENT 'Number of times the schema is found in schema cache during schema inference', `ProfileEvent_SchemaInferenceCacheNumRowsHits` UInt64 COMMENT 'Number of times the number of rows is found in schema cache during count from files', `ProfileEvent_SchemaInferenceCacheMisses` UInt64 COMMENT 'Number of times the requested source is not in schema cache', `ProfileEvent_SchemaInferenceCacheSchemaMisses` UInt64 COMMENT 'Number of times the requested source is in cache but the schema is not in cache during schema inference', `ProfileEvent_SchemaInferenceCacheNumRowsMisses` UInt64 COMMENT 'Number of times the requested source is in cache but the number of rows is not in cache while count from files', `ProfileEvent_SchemaInferenceCacheEvictions` UInt64 COMMENT 'Number of times a schema from cache was evicted due to overflow', `ProfileEvent_SchemaInferenceCacheInvalidations` UInt64 COMMENT 'Number of times a schema in cache became invalid due to changes in data', `ProfileEvent_KeeperPacketsSent` UInt64 COMMENT 'Packets sent by keeper server', `ProfileEvent_KeeperPacketsReceived` UInt64 COMMENT 'Packets received by keeper server', `ProfileEvent_KeeperRequestTotal` UInt64 COMMENT 'Total requests number on keeper server', `ProfileEvent_KeeperLatency` UInt64 COMMENT 'Keeper latency', `ProfileEvent_KeeperTotalElapsedMicroseconds` UInt64 COMMENT 'Keeper total latency for a single request', `ProfileEvent_KeeperProcessElapsedMicroseconds` UInt64 COMMENT 'Keeper commit latency for a single request', `ProfileEvent_KeeperPreprocessElapsedMicroseconds` UInt64 COMMENT 'Keeper preprocessing latency for a single reuquest', `ProfileEvent_KeeperStorageLockWaitMicroseconds` UInt64 COMMENT 'Time spent waiting for acquiring Keeper storage lock', `ProfileEvent_KeeperCommitWaitElapsedMicroseconds` UInt64 COMMENT 'Time spent waiting for certain log to be committed', `ProfileEvent_KeeperBatchMaxCount` UInt64 COMMENT 'Number of times the size of batch was limited by the amount', `ProfileEvent_KeeperBatchMaxTotalSize` UInt64 COMMENT 'Number of times the size of batch was limited by the total bytes size', `ProfileEvent_KeeperCommits` UInt64 COMMENT 'Number of successful commits', `ProfileEvent_KeeperCommitsFailed` UInt64 COMMENT 'Number of failed commits', `ProfileEvent_KeeperSnapshotCreations` UInt64 COMMENT 'Number of snapshots creations', `ProfileEvent_KeeperSnapshotCreationsFailed` UInt64 COMMENT 'Number of failed snapshot creations', `ProfileEvent_KeeperSnapshotApplys` UInt64 COMMENT 'Number of snapshot applying', `ProfileEvent_KeeperSnapshotApplysFailed` UInt64 COMMENT 'Number of failed snapshot applying', `ProfileEvent_KeeperReadSnapshot` UInt64 COMMENT 'Number of snapshot read(serialization)', `ProfileEvent_KeeperSaveSnapshot` UInt64 COMMENT 'Number of snapshot save', `ProfileEvent_KeeperCreateRequest` UInt64 COMMENT 'Number of create requests', `ProfileEvent_KeeperRemoveRequest` UInt64 COMMENT 'Number of remove requests', `ProfileEvent_KeeperSetRequest` UInt64 COMMENT 'Number of set requests', `ProfileEvent_KeeperReconfigRequest` UInt64 COMMENT 'Number of reconfig requests', `ProfileEvent_KeeperCheckRequest` UInt64 COMMENT 'Number of check requests', `ProfileEvent_KeeperMultiRequest` UInt64 COMMENT 'Number of multi requests', `ProfileEvent_KeeperMultiReadRequest` UInt64 COMMENT 'Number of multi read requests', `ProfileEvent_KeeperGetRequest` UInt64 COMMENT 'Number of get requests', `ProfileEvent_KeeperListRequest` UInt64 COMMENT 'Number of list requests', `ProfileEvent_KeeperExistsRequest` UInt64 COMMENT 'Number of exists requests', `ProfileEvent_OverflowBreak` UInt64 COMMENT 'Number of times, data processing was cancelled by query complexity limitation with setting \'*_overflow_mode\' = \'break\' and the result is incomplete.', `ProfileEvent_OverflowThrow` UInt64 COMMENT 'Number of times, data processing was cancelled by query complexity limitation with setting \'*_overflow_mode\' = \'throw\' and exception was thrown.', `ProfileEvent_OverflowAny` UInt64 COMMENT 'Number of times approximate GROUP BY was in effect: when aggregation was performed only on top of first \'max_rows_to_group_by\' unique keys and other keys were ignored due to \'group_by_overflow_mode\' = \'any\'.', `ProfileEvent_S3QueueSetFileProcessingMicroseconds` UInt64 COMMENT 'Time spent to set file as processing', `ProfileEvent_S3QueueSetFileProcessedMicroseconds` UInt64 COMMENT 'Time spent to set file as processed', `ProfileEvent_S3QueueSetFileFailedMicroseconds` UInt64 COMMENT 'Time spent to set file as failed', `ProfileEvent_ObjectStorageQueueFailedFiles` UInt64 COMMENT 'Number of files which failed to be processed', `ProfileEvent_ObjectStorageQueueProcessedFiles` UInt64 COMMENT 'Number of files which were processed', `ProfileEvent_ObjectStorageQueueCleanupMaxSetSizeOrTTLMicroseconds` UInt64 COMMENT 'Time spent to set file as failed', `ProfileEvent_ObjectStorageQueuePullMicroseconds` UInt64 COMMENT 'Time spent to read file data', `ProfileEvent_ObjectStorageQueueLockLocalFileStatusesMicroseconds` UInt64 COMMENT 'Time spent to lock local file statuses', `ProfileEvent_ObjectStorageQueueFailedToBatchSetProcessing` UInt64 COMMENT 'Number of times batched set processing request failed', `ProfileEvent_ObjectStorageQueueTrySetProcessingRequests` UInt64 COMMENT 'The number of times we tried to make set processing request', `ProfileEvent_ObjectStorageQueueTrySetProcessingSucceeded` UInt64 COMMENT 'The number of times we successfully set file as processing', `ProfileEvent_ObjectStorageQueueTrySetProcessingFailed` UInt64 COMMENT 'The number of times we unsuccessfully set file as processing', `ProfileEvent_ObjectStorageQueueListedFiles` UInt64 COMMENT 'Number of listed files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueFilteredFiles` UInt64 COMMENT 'Number of filtered files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueReadFiles` UInt64 COMMENT 'Number of read files (not equal to the number of actually inserted files)', `ProfileEvent_ObjectStorageQueueReadRows` UInt64 COMMENT 'Number of read rows (not equal to the number of actually inserted rows)', `ProfileEvent_ObjectStorageQueueReadBytes` UInt64 COMMENT 'Number of read bytes (not equal to the number of actually inserted bytes)', `ProfileEvent_ObjectStorageQueueExceptionsDuringRead` UInt64 COMMENT 'Number of exceptions during read in S3(Azure)Queue', `ProfileEvent_ObjectStorageQueueExceptionsDuringInsert` UInt64 COMMENT 'Number of exceptions during insert in S3(Azure)Queue', `ProfileEvent_ObjectStorageQueueRemovedObjects` UInt64 COMMENT 'Number of objects removed as part of after_processing = delete', `ProfileEvent_ObjectStorageQueueInsertIterations` UInt64 COMMENT 'Number of insert iterations', `ProfileEvent_ObjectStorageQueueCommitRequests` UInt64 COMMENT 'Number of keeper requests to commit files as either failed or processed', `ProfileEvent_ObjectStorageQueueSuccessfulCommits` UInt64 COMMENT 'Number of successful keeper commits', `ProfileEvent_ObjectStorageQueueUnsuccessfulCommits` UInt64 COMMENT 'Number of unsuccessful keeper commits', `ProfileEvent_ObjectStorageQueueCancelledFiles` UInt64 COMMENT 'Number cancelled files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueProcessedRows` UInt64 COMMENT 'Number of processed rows in StorageS3(Azure)Queue', `ProfileEvent_ServerStartupMilliseconds` UInt64 COMMENT 'Time elapsed from starting server to listening to sockets in milliseconds', `ProfileEvent_IOUringSQEsSubmitted` UInt64 COMMENT 'Total number of io_uring SQEs submitted', `ProfileEvent_IOUringSQEsResubmitsAsync` UInt64 COMMENT 'Total number of asynchronous io_uring SQE resubmits performed', `ProfileEvent_IOUringSQEsResubmitsSync` UInt64 COMMENT 'Total number of synchronous io_uring SQE resubmits performed', `ProfileEvent_IOUringCQEsCompleted` UInt64 COMMENT 'Total number of successfully completed io_uring CQEs', `ProfileEvent_IOUringCQEsFailed` UInt64 COMMENT 'Total number of completed io_uring CQEs with failures', `ProfileEvent_BackupsOpenedForRead` UInt64 COMMENT 'Number of backups opened for reading', `ProfileEvent_BackupsOpenedForWrite` UInt64 COMMENT 'Number of backups opened for writing', `ProfileEvent_BackupsOpenedForUnlock` UInt64 COMMENT 'Number of backups opened for unlocking', `ProfileEvent_BackupReadMetadataMicroseconds` UInt64 COMMENT 'Time spent reading backup metadata from .backup file', `ProfileEvent_BackupWriteMetadataMicroseconds` UInt64 COMMENT 'Time spent writing backup metadata to .backup file', `ProfileEvent_BackupEntriesCollectorMicroseconds` UInt64 COMMENT 'Time spent making backup entries', `ProfileEvent_BackupEntriesCollectorForTablesDataMicroseconds` UInt64 COMMENT 'Time spent making backup entries for tables data', `ProfileEvent_BackupEntriesCollectorRunPostTasksMicroseconds` UInt64 COMMENT 'Time spent running post tasks after making backup entries', `ProfileEvent_BackupPreparingFileInfosMicroseconds` UInt64 COMMENT 'Time spent preparing file infos for backup entries', `ProfileEvent_BackupReadLocalFilesToCalculateChecksums` UInt64 COMMENT 'Number of files read locally to calculate checksums for backup entries', `ProfileEvent_BackupReadLocalBytesToCalculateChecksums` UInt64 COMMENT 'Total size of files read locally to calculate checksums for backup entries', `ProfileEvent_BackupReadRemoteFilesToCalculateChecksums` UInt64 COMMENT 'Number of files read from remote disks to calculate checksums for backup entries', `ProfileEvent_BackupReadRemoteBytesToCalculateChecksums` UInt64 COMMENT 'Total size of files read from remote disks to calculate checksums for backup entries', `ProfileEvent_BackupLockFileReads` UInt64 COMMENT 'How many times the \'.lock\' file was read while making backup', `ProfileEvent_RestorePartsSkippedFiles` UInt64 COMMENT 'Number of files skipped while restoring parts', `ProfileEvent_RestorePartsSkippedBytes` UInt64 COMMENT 'Total size of files skipped while restoring parts', `ProfileEvent_ReadTaskRequestsReceived` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the initiator server side.', `ProfileEvent_MergeTreeReadTaskRequestsReceived` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the initiator server side.', `ProfileEvent_ReadTaskRequestsSent` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `ProfileEvent_MergeTreeReadTaskRequestsSent` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergeTreeAllRangesAnnouncementsSent` UInt64 COMMENT 'The number of announcements sent from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_ReadTaskRequestsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `ProfileEvent_MergeTreeReadTaskRequestsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergeTreeAllRangesAnnouncementsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in sending the announcement from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergerMutatorsGetPartsForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to take data parts snapshot to build ranges from them.', `ProfileEvent_MergerMutatorPrepareRangesForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to prepare parts ranges which can be merged according to merge predicate.', `ProfileEvent_MergerMutatorSelectPartsForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to select parts from ranges which can be merged.', `ProfileEvent_MergerMutatorRangesForMergeCount` UInt64 COMMENT 'Amount of candidate ranges for merge', `ProfileEvent_MergerMutatorPartsInRangesForMergeCount` UInt64 COMMENT 'Amount of candidate parts for merge', `ProfileEvent_MergerMutatorSelectRangePartsCount` UInt64 COMMENT 'Amount of parts in selected range for merge', `ProfileEvent_ConnectionPoolIsFullMicroseconds` UInt64 COMMENT 'Total time spent waiting for a slot in connection pool.', `ProfileEvent_AsyncLoaderWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting for async loader jobs.', `ProfileEvent_DistrCacheServerSwitches` UInt64 COMMENT 'Distributed Cache read buffer event. Number of server switches between distributed cache servers in read/write-through cache', `ProfileEvent_DistrCacheReadMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent reading from distributed cache', `ProfileEvent_DistrCacheFallbackReadMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spend reading from fallback buffer instead of distributed cache', `ProfileEvent_DistrCachePrecomputeRangesMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent to precompute read ranges', `ProfileEvent_DistrCacheNextImplMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spend in ReadBufferFromDistributedCache::nextImpl', `ProfileEvent_DistrCacheStartRangeMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent to start a new read range with distributed cache', `ProfileEvent_DistrCacheIgnoredBytesWhileWaitingProfileEvents` UInt64 COMMENT 'Distributed Cache read buffer event. Ignored bytes while waiting for profile events in distributed cache', `ProfileEvent_DistrCacheRangeChange` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we changed read range because of seek/last_position change', `ProfileEvent_DistrCacheGetResponseMicroseconds` UInt64 COMMENT 'Distributed Cache client event. Time spend to wait for response from distributed cache', `ProfileEvent_DistrCacheReadErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors during read', `ProfileEvent_DistrCacheMakeRequestErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors when making a request', `ProfileEvent_DistrCacheReceiveResponseErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors when receiving response a request', `ProfileEvent_DistrCachePackets` UInt64 COMMENT 'Distributed Cache client event. Total number of packets received from distributed cache', `ProfileEvent_DistrCachePacketsBytes` UInt64 COMMENT 'Distributed Cache client event. The number of bytes in Data packets which were not ignored', `ProfileEvent_DistrCacheUnusedPackets` UInt64 COMMENT 'Distributed Cache client event. Number of skipped unused packets from distributed cache', `ProfileEvent_DistrCacheUnusedPacketsBytes` UInt64 COMMENT 'Distributed Cache client event. The number of bytes in Data packets which were ignored', `ProfileEvent_DistrCacheUnusedPacketsBufferAllocations` UInt64 COMMENT 'Distributed Cache client event. The number of extra buffer allocations in case we could not reuse existing buffer', `ProfileEvent_DistrCacheLockRegistryMicroseconds` UInt64 COMMENT 'Distributed Cache registry event. Time spent to take DistributedCacheRegistry lock', `ProfileEvent_DistrCacheRegistryUpdateMicroseconds` UInt64 COMMENT 'Distributed Cache registry event. Time spent updating distributed cache registry', `ProfileEvent_DistrCacheRegistryUpdates` UInt64 COMMENT 'Distributed Cache registry event. Number of distributed cache registry updates', `ProfileEvent_DistrCacheHashRingRebuilds` UInt64 COMMENT 'Distributed Cache registry event. Number of distributed cache hash ring rebuilds', `ProfileEvent_DistrCacheReadBytesFromCache` UInt64 COMMENT 'Distributed Cache read buffer event. Bytes read from distributed cache', `ProfileEvent_DistrCacheReadBytesFromFallbackBuffer` UInt64 COMMENT 'Distributed Cache read buffer event. Bytes read from fallback buffer', `ProfileEvent_DistrCacheRangeResetBackward` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we reset read range because of seek/last_position change', `ProfileEvent_DistrCacheRangeResetForward` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we reset read range because of seek/last_position change', `ProfileEvent_DistrCacheOpenedConnections` UInt64 COMMENT 'Distributed Cache connection event. The number of open connections to distributed cache', `ProfileEvent_DistrCacheReusedConnections` UInt64 COMMENT 'Distributed Cache connection event. The number of reused connections to distributed cache', `ProfileEvent_DistrCacheOpenedConnectionsBypassingPool` UInt64 COMMENT 'Distributed Cache connection event. The number of open connections to distributed cache bypassing pool', `ProfileEvent_DistrCacheConnectMicroseconds` UInt64 COMMENT 'Distributed Cache connection event. The time spent to connect to distributed cache', `ProfileEvent_DistrCacheConnectAttempts` UInt64 COMMENT 'Distributed Cache connection event. The number of connection attempts to distributed cache', `ProfileEvent_DistrCacheGetClientMicroseconds` UInt64 COMMENT 'Distributed Cache connection event. Time spent getting client for distributed cache', `ProfileEvent_DistrCacheServerProcessRequestMicroseconds` UInt64 COMMENT 'Distributed Cache server event. Time spent processing request on DistributedCache server side', `ProfileEvent_DistrCacheServerStartRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of StartRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerContinueRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of ContinueRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerEndRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of EndRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerAckRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of AckRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerNewS3CachedClients` UInt64 COMMENT 'Distributed Cache server event. The number of new cached s3 clients', `ProfileEvent_DistrCacheServerReusedS3CachedClients` UInt64 COMMENT 'Distributed Cache server event. The number of reused cached s3 clients', `ProfileEvent_LogTest` UInt64 COMMENT 'Number of log messages with level Test', `ProfileEvent_LogTrace` UInt64 COMMENT 'Number of log messages with level Trace', `ProfileEvent_LogDebug` UInt64 COMMENT 'Number of log messages with level Debug', `ProfileEvent_LogInfo` UInt64 COMMENT 'Number of log messages with level Info', `ProfileEvent_LogWarning` UInt64 COMMENT 'Number of log messages with level Warning', `ProfileEvent_LogError` UInt64 COMMENT 'Number of log messages with level Error', `ProfileEvent_LogFatal` UInt64 COMMENT 'Number of log messages with level Fatal', `ProfileEvent_LoggerElapsedNanoseconds` UInt64 COMMENT 'Cumulative time spend in logging', `ProfileEvent_InterfaceHTTPSendBytes` UInt64 COMMENT 'Number of bytes sent through HTTP interfaces', `ProfileEvent_InterfaceHTTPReceiveBytes` UInt64 COMMENT 'Number of bytes received through HTTP interfaces', `ProfileEvent_InterfaceNativeSendBytes` UInt64 COMMENT 'Number of bytes sent through native interfaces', `ProfileEvent_InterfaceNativeReceiveBytes` UInt64 COMMENT 'Number of bytes received through native interfaces', `ProfileEvent_InterfacePrometheusSendBytes` UInt64 COMMENT 'Number of bytes sent through Prometheus interfaces', `ProfileEvent_InterfacePrometheusReceiveBytes` UInt64 COMMENT 'Number of bytes received through Prometheus interfaces', `ProfileEvent_InterfaceInterserverSendBytes` UInt64 COMMENT 'Number of bytes sent through interserver interfaces', `ProfileEvent_InterfaceInterserverReceiveBytes` UInt64 COMMENT 'Number of bytes received through interserver interfaces', `ProfileEvent_InterfaceMySQLSendBytes` UInt64 COMMENT 'Number of bytes sent through MySQL interfaces', `ProfileEvent_InterfaceMySQLReceiveBytes` UInt64 COMMENT 'Number of bytes received through MySQL interfaces', `ProfileEvent_InterfacePostgreSQLSendBytes` UInt64 COMMENT 'Number of bytes sent through PostgreSQL interfaces', `ProfileEvent_InterfacePostgreSQLReceiveBytes` UInt64 COMMENT 'Number of bytes received through PostgreSQL interfaces', `ProfileEvent_ParallelReplicasUsedCount` UInt64 COMMENT 'Number of replicas used to execute a query with task-based parallel replicas', `ProfileEvent_ParallelReplicasAvailableCount` UInt64 COMMENT 'Number of replicas available to execute a query with task-based parallel replicas', `ProfileEvent_ParallelReplicasUnavailableCount` UInt64 COMMENT 'Number of replicas which was chosen, but found to be unavailable during query execution with task-based parallel replicas', `ProfileEvent_SharedMergeTreeVirtualPartsUpdates` UInt64 COMMENT 'Virtual parts update count', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesByLeader` UInt64 COMMENT 'Virtual parts updates by leader', `ProfileEvent_SharedMergeTreeVirtualPartsUpdateMicroseconds` UInt64 COMMENT 'Virtual parts update microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromZooKeeper` UInt64 COMMENT 'Virtual parts updates count from ZooKeeper', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromZooKeeperMicroseconds` UInt64 COMMENT 'Virtual parts updates from ZooKeeper microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesPeerNotFound` UInt64 COMMENT 'Virtual updates from peer failed because no one found', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromPeer` UInt64 COMMENT 'Virtual parts updates count from peer', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromPeerMicroseconds` UInt64 COMMENT 'Virtual parts updates from peer microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesForMergesOrStatus` UInt64 COMMENT 'Virtual parts updates from non-default background job', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesLeaderFailedElection` UInt64 COMMENT 'Virtual parts updates leader election failed', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesLeaderSuccessfulElection` UInt64 COMMENT 'Virtual parts updates leader election successful', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentAttempt` UInt64 COMMENT 'How many times we tried to assign merge or mutation', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentFailedWithNothingToDo` UInt64 COMMENT 'How many times we tried to assign merge or mutation and failed because nothing to merge', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentFailedWithConflict` UInt64 COMMENT 'How many times we tried to assign merge or mutation and failed because of conflict in Keeper', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentSuccessful` UInt64 COMMENT 'How many times we tried to assign merge or mutation', `ProfileEvent_SharedMergeTreeMergePartsMovedToOudated` UInt64 COMMENT 'How many parts moved to oudated directory', `ProfileEvent_SharedMergeTreeMergePartsMovedToCondemned` UInt64 COMMENT 'How many parts moved to condemned directory', `ProfileEvent_SharedMergeTreeOutdatedPartsConfirmationRequest` UInt64 COMMENT 'How many ZooKeeper requests were used to config outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsConfirmationInvocations` UInt64 COMMENT 'How many invocations were made to confirm outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsHTTPRequest` UInt64 COMMENT 'How many HTTP requests were send to confirm outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsHTTPResponse` UInt64 COMMENT 'How many HTTP responses were send to confirm outdated parts', `ProfileEvent_SharedMergeTreeCondemnedPartsKillRequest` UInt64 COMMENT 'How many ZooKeeper requests were used to remove condemned parts', `ProfileEvent_SharedMergeTreeCondemnedPartsLockConfict` UInt64 COMMENT 'How many times we failed to acquite lock because of conflict', `ProfileEvent_SharedMergeTreeCondemnedPartsRemoved` UInt64 COMMENT 'How many condemned parts were removed', `ProfileEvent_KeeperLogsEntryReadFromLatestCache` UInt64 COMMENT 'Number of log entries in Keeper being read from latest logs cache', `ProfileEvent_KeeperLogsEntryReadFromCommitCache` UInt64 COMMENT 'Number of log entries in Keeper being read from commit logs cache', `ProfileEvent_KeeperLogsEntryReadFromFile` UInt64 COMMENT 'Number of log entries in Keeper being read directly from the changelog file', `ProfileEvent_KeeperLogsPrefetchedEntries` UInt64 COMMENT 'Number of log entries in Keeper being prefetched from the changelog file', `ProfileEvent_StorageConnectionsCreated` UInt64 COMMENT 'Number of created connections for storages', `ProfileEvent_StorageConnectionsReused` UInt64 COMMENT 'Number of reused connections for storages', `ProfileEvent_StorageConnectionsReset` UInt64 COMMENT 'Number of reset connections for storages', `ProfileEvent_StorageConnectionsPreserved` UInt64 COMMENT 'Number of preserved connections for storages', `ProfileEvent_StorageConnectionsExpired` UInt64 COMMENT 'Number of expired connections for storages', `ProfileEvent_StorageConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a connection for storage is failed', `ProfileEvent_StorageConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating connections for storages', `ProfileEvent_DiskConnectionsCreated` UInt64 COMMENT 'Number of created connections for disk', `ProfileEvent_DiskConnectionsReused` UInt64 COMMENT 'Number of reused connections for disk', `ProfileEvent_DiskConnectionsReset` UInt64 COMMENT 'Number of reset connections for disk', `ProfileEvent_DiskConnectionsPreserved` UInt64 COMMENT 'Number of preserved connections for disk', `ProfileEvent_DiskConnectionsExpired` UInt64 COMMENT 'Number of expired connections for disk', `ProfileEvent_DiskConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a connection for disk is failed', `ProfileEvent_DiskConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating connections for disk', `ProfileEvent_HTTPConnectionsCreated` UInt64 COMMENT 'Number of created http connections', `ProfileEvent_HTTPConnectionsReused` UInt64 COMMENT 'Number of reused http connections', `ProfileEvent_HTTPConnectionsReset` UInt64 COMMENT 'Number of reset http connections', `ProfileEvent_HTTPConnectionsPreserved` UInt64 COMMENT 'Number of preserved http connections', `ProfileEvent_HTTPConnectionsExpired` UInt64 COMMENT 'Number of expired http connections', `ProfileEvent_HTTPConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a http connection failed', `ProfileEvent_HTTPConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating http connections', `ProfileEvent_AddressesDiscovered` UInt64 COMMENT 'Total count of new addresses in dns resolve results for http connections', `ProfileEvent_AddressesExpired` UInt64 COMMENT 'Total count of expired addresses which is no longer presented in dns resolve results for http connections', `ProfileEvent_AddressesMarkedAsFailed` UInt64 COMMENT 'Total count of addresses which has been marked as faulty due to connection errors for http connections', `ProfileEvent_ReadWriteBufferFromHTTPRequestsSent` UInt64 COMMENT 'Number of HTTP requests sent by ReadWriteBufferFromHTTP', `ProfileEvent_ReadWriteBufferFromHTTPBytes` UInt64 COMMENT 'Total size of payload bytes received and sent by ReadWriteBufferFromHTTP. Doesn\'t include HTTP headers.', `ProfileEvent_ConcurrencyControlSlotsGranted` UInt64 COMMENT 'Number of CPU slot granted according to guarantee of 1 thread per query and for queries with setting \'use_concurrency_control\' = 0', `ProfileEvent_ConcurrencyControlSlotsDelayed` UInt64 COMMENT 'Number of CPU slot not granted initially and required to wait for a free CPU slot', `ProfileEvent_ConcurrencyControlSlotsAcquired` UInt64 COMMENT 'Total number of CPU slot acquired', `ProfileEvent_ConcurrencyControlSlotsAcquiredNonCompeting` UInt64 COMMENT 'Total number of noncompeting CPU slot acquired', `ProfileEvent_ConcurrencyControlQueriesDelayed` UInt64 COMMENT 'Total number of CPU slot allocations (queries) that were required to wait for slots to upscale', `ProfileEvent_SharedDatabaseCatalogFailedToApplyState` UInt64 COMMENT 'Number of failures to apply new state in SharedDatabaseCatalog', `ProfileEvent_SharedDatabaseCatalogStateApplicationMicroseconds` UInt64 COMMENT 'Total time spend on application of new state in SharedDatabaseCatalog', `ProfileEvent_GWPAsanAllocateSuccess` UInt64 COMMENT 'Number of successful allocations done by GWPAsan', `ProfileEvent_GWPAsanAllocateFailed` UInt64 COMMENT 'Number of failed allocations done by GWPAsan (i.e. filled pool)', `ProfileEvent_GWPAsanFree` UInt64 COMMENT 'Number of free operations done by GWPAsan', `ProfileEvent_MemoryWorkerRun` UInt64 COMMENT 'Number of runs done by MemoryWorker in background', `ProfileEvent_MemoryWorkerRunElapsedMicroseconds` UInt64 COMMENT 'Total time spent by MemoryWorker for background work', `ProfileEvent_ParquetFetchWaitTimeMicroseconds` UInt64 COMMENT 'Time of waiting fetching parquet data' ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains history of memory and metric values from table system.events for individual queries, periodically flushed to disk.\n\nIt is safe to truncate or drop this table at any time.' With storage policy, SHOW CREATE TABLE system.query_metric_log is: CREATE TABLE system.query_metric_log ( `query_id` String COMMENT 'Query ID.' CODEC(ZSTD(1)), `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.' CODEC(ZSTD(1)), `event_date` Date COMMENT 'Event date.' CODEC(Delta(2), ZSTD(1)), `event_time` DateTime COMMENT 'Event time.' CODEC(Delta(4), ZSTD(1)), `event_time_microseconds` DateTime64(6) COMMENT 'Event time with microseconds resolution.' CODEC(Delta(4), ZSTD(1)), `memory_usage` UInt64 COMMENT 'Amount of RAM the query uses. It might not include some types of dedicated memory.', `peak_memory_usage` UInt64 COMMENT 'Maximum amount of RAM the query used.', `ProfileEvent_Query` UInt64 COMMENT 'Number of queries to be interpreted and potentially executed. Does not include queries that failed to parse or were rejected due to AST size limits, quota limits or limits on the number of simultaneously running queries. May include internal queries initiated by ClickHouse itself. Does not count subqueries.', `ProfileEvent_SelectQuery` UInt64 COMMENT 'Same as Query, but only for SELECT queries.', `ProfileEvent_InsertQuery` UInt64 COMMENT 'Same as Query, but only for INSERT queries.', `ProfileEvent_InitialQuery` UInt64 COMMENT 'Same as Query, but only counts initial queries (see is_initial_query).', `ProfileEvent_QueriesWithSubqueries` UInt64 COMMENT 'Count queries with all subqueries', `ProfileEvent_SelectQueriesWithSubqueries` UInt64 COMMENT 'Count SELECT queries with all subqueries', `ProfileEvent_InsertQueriesWithSubqueries` UInt64 COMMENT 'Count INSERT queries with all subqueries', `ProfileEvent_SelectQueriesWithPrimaryKeyUsage` UInt64 COMMENT 'Count SELECT queries which use the primary key to evaluate the WHERE condition', `ProfileEvent_AsyncInsertQuery` UInt64 COMMENT 'Same as InsertQuery, but only for asynchronous INSERT queries.', `ProfileEvent_AsyncInsertBytes` UInt64 COMMENT 'Data size in bytes of asynchronous INSERT queries.', `ProfileEvent_AsyncInsertRows` UInt64 COMMENT 'Number of rows inserted by asynchronous INSERT queries.', `ProfileEvent_AsyncInsertCacheHits` UInt64 COMMENT 'Number of times a duplicate hash id has been found in asynchronous INSERT hash id cache.', `ProfileEvent_FailedQuery` UInt64 COMMENT 'Number of failed queries.', `ProfileEvent_FailedSelectQuery` UInt64 COMMENT 'Same as FailedQuery, but only for SELECT queries.', `ProfileEvent_FailedInsertQuery` UInt64 COMMENT 'Same as FailedQuery, but only for INSERT queries.', `ProfileEvent_FailedAsyncInsertQuery` UInt64 COMMENT 'Number of failed ASYNC INSERT queries.', `ProfileEvent_QueryTimeMicroseconds` UInt64 COMMENT 'Total time of all queries.', `ProfileEvent_SelectQueryTimeMicroseconds` UInt64 COMMENT 'Total time of SELECT queries.', `ProfileEvent_InsertQueryTimeMicroseconds` UInt64 COMMENT 'Total time of INSERT queries.', `ProfileEvent_OtherQueryTimeMicroseconds` UInt64 COMMENT 'Total time of queries that are not SELECT or INSERT.', `ProfileEvent_FileOpen` UInt64 COMMENT 'Number of files opened.', `ProfileEvent_Seek` UInt64 COMMENT 'Number of times the \'lseek\' function was called.', `ProfileEvent_ReadBufferFromFileDescriptorRead` UInt64 COMMENT 'Number of reads (read/pread) from a file descriptor. Does not include sockets.', `ProfileEvent_ReadBufferFromFileDescriptorReadFailed` UInt64 COMMENT 'Number of times the read (read/pread) from a file descriptor have failed.', `ProfileEvent_ReadBufferFromFileDescriptorReadBytes` UInt64 COMMENT 'Number of bytes read from file descriptors. If the file is compressed, this will show the compressed data size.', `ProfileEvent_WriteBufferFromFileDescriptorWrite` UInt64 COMMENT 'Number of writes (write/pwrite) to a file descriptor. Does not include sockets.', `ProfileEvent_WriteBufferFromFileDescriptorWriteFailed` UInt64 COMMENT 'Number of times the write (write/pwrite) to a file descriptor have failed.', `ProfileEvent_WriteBufferFromFileDescriptorWriteBytes` UInt64 COMMENT 'Number of bytes written to file descriptors. If the file is compressed, this will show compressed data size.', `ProfileEvent_FileSync` UInt64 COMMENT 'Number of times the F_FULLFSYNC/fsync/fdatasync function was called for files.', `ProfileEvent_DirectorySync` UInt64 COMMENT 'Number of times the F_FULLFSYNC/fsync/fdatasync function was called for directories.', `ProfileEvent_FileSyncElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for F_FULLFSYNC/fsync/fdatasync syscall for files.', `ProfileEvent_DirectorySyncElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for F_FULLFSYNC/fsync/fdatasync syscall for directories.', `ProfileEvent_ReadCompressedBytes` UInt64 COMMENT 'Number of bytes (the number of bytes before decompression) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferBlocks` UInt64 COMMENT 'Number of compressed blocks (the blocks of data that are compressed independent of each other) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferBytes` UInt64 COMMENT 'Number of uncompressed bytes (the number of bytes after decompression) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferChecksumDoesntMatch` UInt64 COMMENT 'Number of times the compressed block checksum did not match.', `ProfileEvent_CompressedReadBufferChecksumDoesntMatchSingleBitMismatch` UInt64 COMMENT 'Number of times a compressed block checksum mismatch was caused by a single-bit difference.', `ProfileEvent_CompressedReadBufferChecksumDoesntMatchMicroseconds` UInt64 COMMENT 'Total time spent detecting bit-flips due to compressed block checksum mismatches.', `ProfileEvent_UncompressedCacheHits` UInt64 COMMENT 'Number of times a block of data has been found in the uncompressed cache (and decompression was avoided).', `ProfileEvent_UncompressedCacheMisses` UInt64 COMMENT 'Number of times a block of data has not been found in the uncompressed cache (and required decompression).', `ProfileEvent_UncompressedCacheWeightLost` UInt64 COMMENT 'Number of bytes evicted from the uncompressed cache.', `ProfileEvent_MMappedFileCacheHits` UInt64 COMMENT 'Number of times a file has been found in the MMap cache (for the \'mmap\' read_method), so we didn\'t have to mmap it again.', `ProfileEvent_MMappedFileCacheMisses` UInt64 COMMENT 'Number of times a file has not been found in the MMap cache (for the \'mmap\' read_method), so we had to mmap it again.', `ProfileEvent_OpenedFileCacheHits` UInt64 COMMENT 'Number of times a file has been found in the opened file cache, so we didn\'t have to open it again.', `ProfileEvent_OpenedFileCacheMisses` UInt64 COMMENT 'Number of times a file has been found in the opened file cache, so we had to open it again.', `ProfileEvent_OpenedFileCacheMicroseconds` UInt64 COMMENT 'Amount of time spent executing OpenedFileCache methods.', `ProfileEvent_AIOWrite` UInt64 COMMENT 'Number of writes with Linux or FreeBSD AIO interface', `ProfileEvent_AIOWriteBytes` UInt64 COMMENT 'Number of bytes written with Linux or FreeBSD AIO interface', `ProfileEvent_AIORead` UInt64 COMMENT 'Number of reads with Linux or FreeBSD AIO interface', `ProfileEvent_AIOReadBytes` UInt64 COMMENT 'Number of bytes read with Linux or FreeBSD AIO interface', `ProfileEvent_IOBufferAllocs` UInt64 COMMENT 'Number of allocations of IO buffers (for ReadBuffer/WriteBuffer).', `ProfileEvent_IOBufferAllocBytes` UInt64 COMMENT 'Number of bytes allocated for IO buffers (for ReadBuffer/WriteBuffer).', `ProfileEvent_ArenaAllocChunks` UInt64 COMMENT 'Number of chunks allocated for memory Arena (used for GROUP BY and similar operations)', `ProfileEvent_ArenaAllocBytes` UInt64 COMMENT 'Number of bytes allocated for memory Arena (used for GROUP BY and similar operations)', `ProfileEvent_FunctionExecute` UInt64 COMMENT 'Number of SQL ordinary function calls (SQL functions are called on per-block basis, so this number represents the number of blocks).', `ProfileEvent_TableFunctionExecute` UInt64 COMMENT 'Number of table function calls.', `ProfileEvent_DefaultImplementationForNullsRows` UInt64 COMMENT 'Number of rows processed by default implementation for nulls in function execution', `ProfileEvent_DefaultImplementationForNullsRowsWithNulls` UInt64 COMMENT 'Number of rows which contain null values processed by default implementation for nulls in function execution', `ProfileEvent_MarkCacheHits` UInt64 COMMENT 'Number of times an entry has been found in the mark cache, so we didn\'t have to load a mark file.', `ProfileEvent_MarkCacheMisses` UInt64 COMMENT 'Number of times an entry has not been found in the mark cache, so we had to load a mark file in memory, which is a costly operation, adding to query latency.', `ProfileEvent_PrimaryIndexCacheHits` UInt64 COMMENT 'Number of times an entry has been found in the primary index cache, so we didn\'t have to load a index file.', `ProfileEvent_PrimaryIndexCacheMisses` UInt64 COMMENT 'Number of times an entry has not been found in the primary index cache, so we had to load a index file in memory, which is a costly operation, adding to query latency.', `ProfileEvent_SkippingIndexCacheHits` UInt64 COMMENT 'Number of times an index granule has been found in the skipping index cache.', `ProfileEvent_SkippingIndexCacheMisses` UInt64 COMMENT 'Number of times an index granule has not been found in the skipping index cache and had to be read from disk.', `ProfileEvent_SkippingIndexCacheWeightLost` UInt64 COMMENT 'Approximate number of bytes evicted from the secondary index cache.', `ProfileEvent_QueryCacheHits` UInt64 COMMENT 'Number of times a query result has been found in the query cache (and query computation was avoided). Only updated for SELECT queries with SETTING use_query_cache = 1.', `ProfileEvent_QueryCacheMisses` UInt64 COMMENT 'Number of times a query result has not been found in the query cache (and required query computation). Only updated for SELECT queries with SETTING use_query_cache = 1.', `ProfileEvent_PageCacheChunkMisses` UInt64 COMMENT 'Number of times a chunk has not been found in the userspace page cache.', `ProfileEvent_PageCacheChunkShared` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, already in use by another thread.', `ProfileEvent_PageCacheChunkDataHits` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, with all pages intact.', `ProfileEvent_PageCacheChunkDataPartialHits` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, but some of its pages were evicted by the OS.', `ProfileEvent_PageCacheChunkDataMisses` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, but all its pages were evicted by the OS.', `ProfileEvent_PageCacheBytesUnpinnedRoundedToPages` UInt64 COMMENT 'Total size of populated pages in chunks that became evictable in PageCache. Rounded up to whole pages.', `ProfileEvent_PageCacheBytesUnpinnedRoundedToHugePages` UInt64 COMMENT 'See PageCacheBytesUnpinnedRoundedToPages, but rounded to huge pages. Use the ratio between the two as a measure of memory waste from using huge pages.', `ProfileEvent_CreatedReadBufferOrdinary` UInt64 COMMENT 'Number of times ordinary read buffer was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferDirectIO` UInt64 COMMENT 'Number of times a read buffer with O_DIRECT was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferDirectIOFailed` UInt64 COMMENT 'Number of times a read buffer with O_DIRECT was attempted to be created for reading data (while choosing among other read methods), but the OS did not allow it (due to lack of filesystem support or other reasons) and we fallen back to the ordinary reading method.', `ProfileEvent_CreatedReadBufferMMap` UInt64 COMMENT 'Number of times a read buffer using \'mmap\' was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferMMapFailed` UInt64 COMMENT 'Number of times a read buffer with \'mmap\' was attempted to be created for reading data (while choosing among other read methods), but the OS did not allow it (due to lack of filesystem support or other reasons) and we fallen back to the ordinary reading method.', `ProfileEvent_DiskReadElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for read syscall. This include reads from page cache.', `ProfileEvent_DiskWriteElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for write syscall. This include writes to page cache.', `ProfileEvent_NetworkReceiveElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for data to receive or receiving data from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkSendElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for data to send to network or sending data to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkReceiveBytes` UInt64 COMMENT 'Total number of bytes received from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkSendBytes` UInt64 COMMENT 'Total number of bytes send to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_GlobalThreadPoolExpansions` UInt64 COMMENT 'Counts the total number of times new threads have been added to the global thread pool. This metric indicates the frequency of expansions in the global thread pool to accommodate increased processing demands.', `ProfileEvent_GlobalThreadPoolShrinks` UInt64 COMMENT 'Counts the total number of times the global thread pool has shrunk by removing threads. This occurs when the number of idle threads exceeds max_thread_pool_free_size, indicating adjustments in the global thread pool size in response to decreased thread utilization.', `ProfileEvent_GlobalThreadPoolThreadCreationMicroseconds` UInt64 COMMENT 'Total time spent waiting for new threads to start.', `ProfileEvent_GlobalThreadPoolLockWaitMicroseconds` UInt64 COMMENT 'Total time threads have spent waiting for locks in the global thread pool.', `ProfileEvent_GlobalThreadPoolJobs` UInt64 COMMENT 'Counts the number of jobs that have been pushed to the global thread pool.', `ProfileEvent_GlobalThreadPoolJobWaitTimeMicroseconds` UInt64 COMMENT 'Measures the elapsed time from when a job is scheduled in the thread pool to when it is picked up for execution by a worker thread. This metric helps identify delays in job processing, indicating the responsiveness of the thread pool to new tasks.', `ProfileEvent_LocalThreadPoolExpansions` UInt64 COMMENT 'Counts the total number of times threads have been borrowed from the global thread pool to expand local thread pools.', `ProfileEvent_LocalThreadPoolShrinks` UInt64 COMMENT 'Counts the total number of times threads have been returned to the global thread pool from local thread pools.', `ProfileEvent_LocalThreadPoolThreadCreationMicroseconds` UInt64 COMMENT 'Total time local thread pools have spent waiting to borrow a thread from the global pool.', `ProfileEvent_LocalThreadPoolLockWaitMicroseconds` UInt64 COMMENT 'Total time threads have spent waiting for locks in the local thread pools.', `ProfileEvent_LocalThreadPoolJobs` UInt64 COMMENT 'Counts the number of jobs that have been pushed to the local thread pools.', `ProfileEvent_LocalThreadPoolBusyMicroseconds` UInt64 COMMENT 'Total time threads have spent executing the actual work.', `ProfileEvent_LocalThreadPoolJobWaitTimeMicroseconds` UInt64 COMMENT 'Measures the elapsed time from when a job is scheduled in the thread pool to when it is picked up for execution by a worker thread. This metric helps identify delays in job processing, indicating the responsiveness of the thread pool to new tasks.', `ProfileEvent_DiskS3GetRequestThrottlerCount` UInt64 COMMENT 'Number of DiskS3 GET and SELECT requests passed through throttler.', `ProfileEvent_DiskS3GetRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform DiskS3 GET and SELECT request throttling.', `ProfileEvent_DiskS3PutRequestThrottlerCount` UInt64 COMMENT 'Number of DiskS3 PUT, COPY, POST and LIST requests passed through throttler.', `ProfileEvent_DiskS3PutRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform DiskS3 PUT, COPY, POST and LIST request throttling.', `ProfileEvent_S3GetRequestThrottlerCount` UInt64 COMMENT 'Number of S3 GET and SELECT requests passed through throttler.', `ProfileEvent_S3GetRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform S3 GET and SELECT request throttling.', `ProfileEvent_S3PutRequestThrottlerCount` UInt64 COMMENT 'Number of S3 PUT, COPY, POST and LIST requests passed through throttler.', `ProfileEvent_S3PutRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform S3 PUT, COPY, POST and LIST request throttling.', `ProfileEvent_RemoteReadThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_remote_read_network_bandwidth_for_server\'/\'max_remote_read_network_bandwidth\' throttler.', `ProfileEvent_RemoteReadThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_remote_read_network_bandwidth_for_server\'/\'max_remote_read_network_bandwidth\' throttling.', `ProfileEvent_RemoteWriteThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_remote_write_network_bandwidth_for_server\'/\'max_remote_write_network_bandwidth\' throttler.', `ProfileEvent_RemoteWriteThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_remote_write_network_bandwidth_for_server\'/\'max_remote_write_network_bandwidth\' throttling.', `ProfileEvent_LocalReadThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_local_read_bandwidth_for_server\'/\'max_local_read_bandwidth\' throttler.', `ProfileEvent_LocalReadThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_local_read_bandwidth_for_server\'/\'max_local_read_bandwidth\' throttling.', `ProfileEvent_LocalWriteThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_local_write_bandwidth_for_server\'/\'max_local_write_bandwidth\' throttler.', `ProfileEvent_LocalWriteThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_local_write_bandwidth_for_server\'/\'max_local_write_bandwidth\' throttling.', `ProfileEvent_ThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform all throttling settings.', `ProfileEvent_ReadTasksWithAppliedMutationsOnFly` UInt64 COMMENT 'Total number of parts for which there was any mutation applied on fly', `ProfileEvent_MutationsAppliedOnFlyInAllReadTasks` UInt64 COMMENT 'The sum of number of applied mutations on-fly for part among all read parts', `ProfileEvent_SchedulerIOReadRequests` UInt64 COMMENT 'Resource requests passed through scheduler for IO reads.', `ProfileEvent_SchedulerIOReadBytes` UInt64 COMMENT 'Bytes passed through scheduler for IO reads.', `ProfileEvent_SchedulerIOReadWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting on resource requests for IO reads.', `ProfileEvent_SchedulerIOWriteRequests` UInt64 COMMENT 'Resource requests passed through scheduler for IO writes.', `ProfileEvent_SchedulerIOWriteBytes` UInt64 COMMENT 'Bytes passed through scheduler for IO writes.', `ProfileEvent_SchedulerIOWriteWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting on resource requests for IO writes.', `ProfileEvent_QueryMaskingRulesMatch` UInt64 COMMENT 'Number of times query masking rules was successfully matched.', `ProfileEvent_ReplicatedPartFetches` UInt64 COMMENT 'Number of times a data part was downloaded from replica of a ReplicatedMergeTree table.', `ProfileEvent_ReplicatedPartFailedFetches` UInt64 COMMENT 'Number of times a data part was failed to download from replica of a ReplicatedMergeTree table.', `ProfileEvent_ObsoleteReplicatedParts` UInt64 COMMENT 'Number of times a data part was covered by another data part that has been fetched from a replica (so, we have marked a covered data part as obsolete and no longer needed).', `ProfileEvent_ReplicatedPartMerges` UInt64 COMMENT 'Number of times data parts of ReplicatedMergeTree tables were successfully merged.', `ProfileEvent_ReplicatedPartFetchesOfMerged` UInt64 COMMENT 'Number of times we prefer to download already merged part from replica of ReplicatedMergeTree table instead of performing a merge ourself (usually we prefer doing a merge ourself to save network traffic). This happens when we have not all source parts to perform a merge or when the data part is old enough.', `ProfileEvent_ReplicatedPartMutations` UInt64 COMMENT 'Number of times data parts of ReplicatedMergeTree tables were successfully mutated.', `ProfileEvent_ReplicatedPartChecks` UInt64 COMMENT 'Number of times we had to perform advanced search for a data part on replicas or to clarify the need of an existing data part.', `ProfileEvent_ReplicatedPartChecksFailed` UInt64 COMMENT 'Number of times the advanced search for a data part on replicas did not give result or when unexpected part has been found and moved away.', `ProfileEvent_ReplicatedDataLoss` UInt64 COMMENT 'Number of times a data part that we wanted doesn\'t exist on any replica (even on replicas that are offline right now). That data parts are definitely lost. This is normal due to asynchronous replication (if quorum inserts were not enabled), when the replica on which the data part was written was failed and when it became online after fail it doesn\'t contain that data part.', `ProfileEvent_ReplicatedCoveredPartsInZooKeeperOnStart` UInt64 COMMENT 'For debugging purposes. Number of parts in ZooKeeper that have a covering part, but doesn\'t exist on disk. Checked on server start.', `ProfileEvent_InsertedRows` UInt64 COMMENT 'Number of rows INSERTed to all tables.', `ProfileEvent_InsertedBytes` UInt64 COMMENT 'Number of bytes (uncompressed; for columns as they stored in memory) INSERTed to all tables.', `ProfileEvent_DelayedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a MergeTree table was throttled due to high number of active data parts for partition.', `ProfileEvent_RejectedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a MergeTree table was rejected with \'Too many parts\' exception due to high number of active data parts for partition.', `ProfileEvent_DelayedInsertsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the INSERT of a block to a MergeTree table was throttled due to high number of active data parts for partition.', `ProfileEvent_DelayedMutations` UInt64 COMMENT 'Number of times the mutation of a MergeTree table was throttled due to high number of unfinished mutations for table.', `ProfileEvent_RejectedMutations` UInt64 COMMENT 'Number of times the mutation of a MergeTree table was rejected with \'Too many mutations\' exception due to high number of unfinished mutations for table.', `ProfileEvent_DelayedMutationsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the mutation of a MergeTree table was throttled due to high number of unfinished mutations for table.', `ProfileEvent_DistributedDelayedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a Distributed table was throttled due to high number of pending bytes.', `ProfileEvent_DistributedRejectedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a Distributed table was rejected with \'Too many bytes\' exception due to high number of pending bytes.', `ProfileEvent_DistributedDelayedInsertsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the INSERT of a block to a Distributed table was throttled due to high number of pending bytes.', `ProfileEvent_DuplicatedInsertedBlocks` UInt64 COMMENT 'Number of times the INSERTed block to a ReplicatedMergeTree table was deduplicated.', `ProfileEvent_ZooKeeperInit` UInt64 COMMENT 'Number of times connection with ZooKeeper has been established.', `ProfileEvent_ZooKeeperTransactions` UInt64 COMMENT 'Number of ZooKeeper operations, which include both read and write operations as well as multi-transactions.', `ProfileEvent_ZooKeeperList` UInt64 COMMENT 'Number of \'list\' (getChildren) requests to ZooKeeper.', `ProfileEvent_ZooKeeperCreate` UInt64 COMMENT 'Number of \'create\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperRemove` UInt64 COMMENT 'Number of \'remove\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperExists` UInt64 COMMENT 'Number of \'exists\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperGet` UInt64 COMMENT 'Number of \'get\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperSet` UInt64 COMMENT 'Number of \'set\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperMulti` UInt64 COMMENT 'Number of \'multi\' requests to ZooKeeper (compound transactions).', `ProfileEvent_ZooKeeperCheck` UInt64 COMMENT 'Number of \'check\' requests to ZooKeeper. Usually they don\'t make sense in isolation, only as part of a complex transaction.', `ProfileEvent_ZooKeeperSync` UInt64 COMMENT 'Number of \'sync\' requests to ZooKeeper. These requests are rarely needed or usable.', `ProfileEvent_ZooKeeperReconfig` UInt64 COMMENT 'Number of \'reconfig\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperClose` UInt64 COMMENT 'Number of times connection with ZooKeeper has been closed voluntary.', `ProfileEvent_ZooKeeperWatchResponse` UInt64 COMMENT 'Number of times watch notification has been received from ZooKeeper.', `ProfileEvent_ZooKeeperUserExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper related to the data (no node, bad version or similar).', `ProfileEvent_ZooKeeperHardwareExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper related to network (connection loss or similar).', `ProfileEvent_ZooKeeperOtherExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper other than ZooKeeperUserExceptions and ZooKeeperHardwareExceptions.', `ProfileEvent_ZooKeeperWaitMicroseconds` UInt64 COMMENT 'Number of microseconds spent waiting for responses from ZooKeeper after creating a request, summed across all the requesting threads.', `ProfileEvent_ZooKeeperBytesSent` UInt64 COMMENT 'Number of bytes send over network while communicating with ZooKeeper.', `ProfileEvent_ZooKeeperBytesReceived` UInt64 COMMENT 'Number of bytes received over network while communicating with ZooKeeper.', `ProfileEvent_DistributedConnectionTries` UInt64 COMMENT 'Total count of distributed connection attempts.', `ProfileEvent_DistributedConnectionUsable` UInt64 COMMENT 'Total count of successful distributed connections to a usable server (with required table, but maybe stale).', `ProfileEvent_DistributedConnectionFailTry` UInt64 COMMENT 'Total count when distributed connection fails with retry.', `ProfileEvent_DistributedConnectionMissingTable` UInt64 COMMENT 'Number of times we rejected a replica from a distributed query, because it did not contain a table needed for the query.', `ProfileEvent_DistributedConnectionStaleReplica` UInt64 COMMENT 'Number of times we rejected a replica from a distributed query, because some table needed for a query had replication lag higher than the configured threshold.', `ProfileEvent_DistributedConnectionSkipReadOnlyReplica` UInt64 COMMENT 'Number of replicas skipped during INSERT into Distributed table due to replicas being read-only', `ProfileEvent_DistributedConnectionFailAtAll` UInt64 COMMENT 'Total count when distributed connection fails after all retries finished.', `ProfileEvent_HedgedRequestsChangeReplica` UInt64 COMMENT 'Total count when timeout for changing replica expired in hedged requests.', `ProfileEvent_SuspendSendingQueryToShard` UInt64 COMMENT 'Total count when sending query to shard was suspended when async_query_sending_for_remote is enabled.', `ProfileEvent_CompileFunction` UInt64 COMMENT 'Number of times a compilation of generated LLVM code (to create fused function for complex expressions) was initiated.', `ProfileEvent_CompiledFunctionExecute` UInt64 COMMENT 'Number of times a compiled function was executed.', `ProfileEvent_CompileExpressionsMicroseconds` UInt64 COMMENT 'Total time spent for compilation of expressions to LLVM code.', `ProfileEvent_CompileExpressionsBytes` UInt64 COMMENT 'Number of bytes used for expressions compilation.', `ProfileEvent_ExecuteShellCommand` UInt64 COMMENT 'Number of shell command executions.', `ProfileEvent_ExternalProcessingCompressedBytesTotal` UInt64 COMMENT 'Number of compressed bytes written by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalProcessingUncompressedBytesTotal` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalProcessingFilesTotal` UInt64 COMMENT 'Number of files used by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalSortWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for sorting in external memory.', `ProfileEvent_ExternalSortMerge` UInt64 COMMENT 'Number of times temporary files were merged for sorting in external memory.', `ProfileEvent_ExternalSortCompressedBytes` UInt64 COMMENT 'Number of compressed bytes written for sorting in external memory.', `ProfileEvent_ExternalSortUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written for sorting in external memory.', `ProfileEvent_ExternalAggregationWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for aggregation in external memory.', `ProfileEvent_ExternalAggregationMerge` UInt64 COMMENT 'Number of times temporary files were merged for aggregation in external memory.', `ProfileEvent_ExternalAggregationCompressedBytes` UInt64 COMMENT 'Number of bytes written to disk for aggregation in external memory.', `ProfileEvent_ExternalAggregationUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written to disk for aggregation in external memory.', `ProfileEvent_ExternalJoinWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for JOIN in external memory.', `ProfileEvent_ExternalJoinMerge` UInt64 COMMENT 'Number of times temporary files were merged for JOIN in external memory.', `ProfileEvent_ExternalJoinCompressedBytes` UInt64 COMMENT 'Number of compressed bytes written for JOIN in external memory.', `ProfileEvent_ExternalJoinUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written for JOIN in external memory.', `ProfileEvent_IcebergPartitionPrunnedFiles` UInt64 COMMENT 'Number of skipped files during Iceberg partition pruning', `ProfileEvent_JoinBuildTableRowCount` UInt64 COMMENT 'Total number of rows in the build table for a JOIN operation.', `ProfileEvent_JoinProbeTableRowCount` UInt64 COMMENT 'Total number of rows in the probe table for a JOIN operation.', `ProfileEvent_JoinResultRowCount` UInt64 COMMENT 'Total number of rows in the result of a JOIN operation.', `ProfileEvent_SlowRead` UInt64 COMMENT 'Number of reads from a file that were slow. This indicate system overload. Thresholds are controlled by read_backoff_* settings.', `ProfileEvent_ReadBackoff` UInt64 COMMENT 'Number of times the number of query processing threads was lowered due to slow reads.', `ProfileEvent_ReplicaPartialShutdown` UInt64 COMMENT 'How many times Replicated table has to deinitialize its state due to session expiration in ZooKeeper. The state is reinitialized every time when ZooKeeper is available again.', `ProfileEvent_SelectedParts` UInt64 COMMENT 'Number of data parts selected to read from a MergeTree table.', `ProfileEvent_SelectedPartsTotal` UInt64 COMMENT 'Number of total data parts before selecting which ones to read from a MergeTree table.', `ProfileEvent_SelectedRanges` UInt64 COMMENT 'Number of (non-adjacent) ranges in all data parts selected to read from a MergeTree table.', `ProfileEvent_SelectedMarks` UInt64 COMMENT 'Number of marks (index granules) selected to read from a MergeTree table.', `ProfileEvent_SelectedMarksTotal` UInt64 COMMENT 'Number of total marks (index granules) before selecting which ones to read from a MergeTree table.', `ProfileEvent_SelectedRows` UInt64 COMMENT 'Number of rows SELECTed from all tables.', `ProfileEvent_SelectedBytes` UInt64 COMMENT 'Number of bytes (uncompressed; for columns as they stored in memory) SELECTed from all tables.', `ProfileEvent_RowsReadByMainReader` UInt64 COMMENT 'Number of rows read from MergeTree tables by the main reader (after PREWHERE step).', `ProfileEvent_RowsReadByPrewhereReaders` UInt64 COMMENT 'Number of rows read from MergeTree tables (in total) by prewhere readers.', `ProfileEvent_LoadedDataParts` UInt64 COMMENT 'Number of data parts loaded by MergeTree tables during initialization.', `ProfileEvent_LoadedDataPartsMicroseconds` UInt64 COMMENT 'Microseconds spent by MergeTree tables for loading data parts during initialization.', `ProfileEvent_WaitMarksLoadMicroseconds` UInt64 COMMENT 'Time spent loading marks', `ProfileEvent_BackgroundLoadingMarksTasks` UInt64 COMMENT 'Number of background tasks for loading marks', `ProfileEvent_LoadingMarksTasksCanceled` UInt64 COMMENT 'Number of times background tasks for loading marks were canceled', `ProfileEvent_LoadedMarksFiles` UInt64 COMMENT 'Number of mark files loaded.', `ProfileEvent_LoadedMarksCount` UInt64 COMMENT 'Number of marks loaded (total across columns).', `ProfileEvent_LoadedMarksMemoryBytes` UInt64 COMMENT 'Size of in-memory representations of loaded marks.', `ProfileEvent_LoadedPrimaryIndexFiles` UInt64 COMMENT 'Number of primary index files loaded.', `ProfileEvent_LoadedPrimaryIndexRows` UInt64 COMMENT 'Number of rows of primary key loaded.', `ProfileEvent_LoadedPrimaryIndexBytes` UInt64 COMMENT 'Number of rows of primary key loaded.', `ProfileEvent_Merge` UInt64 COMMENT 'Number of launched background merges.', `ProfileEvent_MergeSourceParts` UInt64 COMMENT 'Number of source parts scheduled for merges.', `ProfileEvent_MergedRows` UInt64 COMMENT 'Rows read for background merges. This is the number of rows before merge.', `ProfileEvent_MergedColumns` UInt64 COMMENT 'Number of columns merged during the horizontal stage of merges.', `ProfileEvent_GatheredColumns` UInt64 COMMENT 'Number of columns gathered during the vertical stage of merges.', `ProfileEvent_MergedUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) that was read for background merges. This is the number before merge.', `ProfileEvent_MergeTotalMilliseconds` UInt64 COMMENT 'Total time spent for background merges', `ProfileEvent_MergeExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of background merges', `ProfileEvent_MergeHorizontalStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for horizontal stage of background merges', `ProfileEvent_MergeHorizontalStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of horizontal stage of background merges', `ProfileEvent_MergeVerticalStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for vertical stage of background merges', `ProfileEvent_MergeVerticalStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of vertical stage of background merges', `ProfileEvent_MergeProjectionStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for projection stage of background merges', `ProfileEvent_MergeProjectionStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of projection stage of background merges', `ProfileEvent_MergePrewarmStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for prewarm stage of background merges', `ProfileEvent_MergePrewarmStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of prewarm stage of background merges', `ProfileEvent_MergingSortedMilliseconds` UInt64 COMMENT 'Total time spent while merging sorted columns', `ProfileEvent_AggregatingSortedMilliseconds` UInt64 COMMENT 'Total time spent while aggregating sorted columns', `ProfileEvent_CollapsingSortedMilliseconds` UInt64 COMMENT 'Total time spent while collapsing sorted columns', `ProfileEvent_ReplacingSortedMilliseconds` UInt64 COMMENT 'Total time spent while replacing sorted columns', `ProfileEvent_SummingSortedMilliseconds` UInt64 COMMENT 'Total time spent while summing sorted columns', `ProfileEvent_VersionedCollapsingSortedMilliseconds` UInt64 COMMENT 'Total time spent while version collapsing sorted columns', `ProfileEvent_GatheringColumnMilliseconds` UInt64 COMMENT 'Total time spent while gathering columns for vertical merge', `ProfileEvent_MutationTotalParts` UInt64 COMMENT 'Number of total parts for which mutations tried to be applied', `ProfileEvent_MutationUntouchedParts` UInt64 COMMENT 'Number of total parts for which mutations tried to be applied but which was completely skipped according to predicate', `ProfileEvent_MutatedRows` UInt64 COMMENT 'Rows read for mutations. This is the number of rows before mutation', `ProfileEvent_MutatedUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) that was read for mutations. This is the number before mutation.', `ProfileEvent_MutationTotalMilliseconds` UInt64 COMMENT 'Total time spent for mutations.', `ProfileEvent_MutationExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of mutations.', `ProfileEvent_MutationAllPartColumns` UInt64 COMMENT 'Number of times when task to mutate all columns in part was created', `ProfileEvent_MutationSomePartColumns` UInt64 COMMENT 'Number of times when task to mutate some columns in part was created', `ProfileEvent_MutateTaskProjectionsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating projections in mutations', `ProfileEvent_MergeTreeDataWriterRows` UInt64 COMMENT 'Number of rows INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterCompressedBytes` UInt64 COMMENT 'Bytes written to filesystem for data INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterBlocks` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables. Each block forms a data part of level zero.', `ProfileEvent_MergeTreeDataWriterBlocksAlreadySorted` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables that appeared to be already sorted.', `ProfileEvent_MergeTreeDataWriterSkipIndicesCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating skip indices', `ProfileEvent_MergeTreeDataWriterStatisticsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating statistics', `ProfileEvent_MergeTreeDataWriterSortingBlocksMicroseconds` UInt64 COMMENT 'Time spent sorting blocks', `ProfileEvent_MergeTreeDataWriterMergingBlocksMicroseconds` UInt64 COMMENT 'Time spent merging input blocks (for special MergeTree engines)', `ProfileEvent_MergeTreeDataWriterProjectionsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating projections', `ProfileEvent_MergeTreeDataProjectionWriterSortingBlocksMicroseconds` UInt64 COMMENT 'Time spent sorting blocks (for projection it might be a key different from table\'s sorting key)', `ProfileEvent_MergeTreeDataProjectionWriterMergingBlocksMicroseconds` UInt64 COMMENT 'Time spent merging blocks', `ProfileEvent_InsertedWideParts` UInt64 COMMENT 'Number of parts inserted in Wide format.', `ProfileEvent_InsertedCompactParts` UInt64 COMMENT 'Number of parts inserted in Compact format.', `ProfileEvent_MergedIntoWideParts` UInt64 COMMENT 'Number of parts merged into Wide format.', `ProfileEvent_MergedIntoCompactParts` UInt64 COMMENT 'Number of parts merged into Compact format.', `ProfileEvent_MergeTreeDataProjectionWriterRows` UInt64 COMMENT 'Number of rows INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterCompressedBytes` UInt64 COMMENT 'Bytes written to filesystem for data INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterBlocks` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables projection. Each block forms a data part of level zero.', `ProfileEvent_MergeTreeDataProjectionWriterBlocksAlreadySorted` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables projection that appeared to be already sorted.', `ProfileEvent_CannotRemoveEphemeralNode` UInt64 COMMENT 'Number of times an error happened while trying to remove ephemeral node. This is not an issue, because our implementation of ZooKeeper library guarantee that the session will expire and the node will be removed.', `ProfileEvent_RegexpWithMultipleNeedlesCreated` UInt64 COMMENT 'Regular expressions with multiple needles (VectorScan library) compiled.', `ProfileEvent_RegexpWithMultipleNeedlesGlobalCacheHit` UInt64 COMMENT 'Number of times we fetched compiled regular expression with multiple needles (VectorScan library) from the global cache.', `ProfileEvent_RegexpWithMultipleNeedlesGlobalCacheMiss` UInt64 COMMENT 'Number of times we failed to fetch compiled regular expression with multiple needles (VectorScan library) from the global cache.', `ProfileEvent_RegexpLocalCacheHit` UInt64 COMMENT 'Number of times we fetched compiled regular expression from a local cache.', `ProfileEvent_RegexpLocalCacheMiss` UInt64 COMMENT 'Number of times we failed to fetch compiled regular expression from a local cache.', `ProfileEvent_ContextLock` UInt64 COMMENT 'Number of times the lock of Context was acquired or tried to acquire. This is global lock.', `ProfileEvent_ContextLockWaitMicroseconds` UInt64 COMMENT 'Context lock wait time in microseconds', `ProfileEvent_StorageBufferFlush` UInt64 COMMENT 'Number of times a buffer in a \'Buffer\' table was flushed.', `ProfileEvent_StorageBufferErrorOnFlush` UInt64 COMMENT 'Number of times a buffer in the \'Buffer\' table has not been able to flush due to error writing in the destination table.', `ProfileEvent_StorageBufferPassedAllMinThresholds` UInt64 COMMENT 'Number of times a criteria on min thresholds has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedTimeMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max time threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedRowsMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max rows threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedBytesMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max bytes threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedTimeFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on time has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferPassedRowsFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on rows has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferPassedBytesFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on bytes has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferLayerLockReadersWaitMilliseconds` UInt64 COMMENT 'Time for waiting for Buffer layer during reading.', `ProfileEvent_StorageBufferLayerLockWritersWaitMilliseconds` UInt64 COMMENT 'Time for waiting free Buffer layer to write to (can be used to tune Buffer layers).', `ProfileEvent_SystemLogErrorOnFlush` UInt64 COMMENT 'Number of times any of the system logs have failed to flush to the corresponding system table. Attempts to flush are repeated.', `ProfileEvent_DictCacheKeysRequested` UInt64 COMMENT 'Number of keys requested from the data source for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheKeysRequestedMiss` UInt64 COMMENT 'Number of keys requested from the data source for dictionaries of \'cache\' types but not found in the data source.', `ProfileEvent_DictCacheKeysRequestedFound` UInt64 COMMENT 'Number of keys requested from the data source for dictionaries of \'cache\' types and found in the data source.', `ProfileEvent_DictCacheKeysExpired` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and found in the cache but they were obsolete.', `ProfileEvent_DictCacheKeysNotFound` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and not found.', `ProfileEvent_DictCacheKeysHit` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and found in the cache.', `ProfileEvent_DictCacheRequestTimeNs` UInt64 COMMENT 'Number of nanoseconds spend in querying the external data sources for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheRequests` UInt64 COMMENT 'Number of bulk requests to the external data sources for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheLockWriteNs` UInt64 COMMENT 'Number of nanoseconds spend in waiting for write lock to update the data for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheLockReadNs` UInt64 COMMENT 'Number of nanoseconds spend in waiting for read lock to lookup the data for the dictionaries of \'cache\' types.', `ProfileEvent_DistributedSyncInsertionTimeoutExceeded` UInt64 COMMENT 'A timeout has exceeded while waiting for shards during synchronous insertion into a Distributed table (with \'distributed_foreground_insert\' = 1)', `ProfileEvent_DistributedAsyncInsertionFailures` UInt64 COMMENT 'Number of failures for asynchronous insertion into a Distributed table (with \'distributed_foreground_insert\' = 0)', `ProfileEvent_DataAfterMergeDiffersFromReplica` UInt64 COMMENT '\nNumber of times data after merge is not byte-identical to the data on another replicas. There could be several reasons:\n1. Using newer version of compression library after server update.\n2. Using another compression method.\n3. Non-deterministic compression algorithm (highly unlikely).\n4. Non-deterministic merge algorithm due to logical error in code.\n5. Data corruption in memory due to bug in code.\n6. Data corruption in memory due to hardware issue.\n7. Manual modification of source data after server startup.\n8. Manual modification of checksums stored in ZooKeeper.\n9. Part format related settings like \'enable_mixed_granularity_parts\' are different on different replicas.\nThe server successfully detected this situation and will download merged part from the replica to force the byte-identical result.\n', `ProfileEvent_DataAfterMutationDiffersFromReplica` UInt64 COMMENT 'Number of times data after mutation is not byte-identical to the data on other replicas. In addition to the reasons described in \'DataAfterMergeDiffersFromReplica\', it is also possible due to non-deterministic mutation.', `ProfileEvent_PolygonsAddedToPool` UInt64 COMMENT 'A polygon has been added to the cache (pool) for the \'pointInPolygon\' function.', `ProfileEvent_PolygonsInPoolAllocatedBytes` UInt64 COMMENT 'The number of bytes for polygons added to the cache (pool) for the \'pointInPolygon\' function.', `ProfileEvent_USearchAddCount` UInt64 COMMENT 'Number of vectors added to usearch indexes.', `ProfileEvent_USearchAddVisitedMembers` UInt64 COMMENT 'Number of nodes visited when adding vectors to usearch indexes.', `ProfileEvent_USearchAddComputedDistances` UInt64 COMMENT 'Number of times distance was computed when adding vectors to usearch indexes.', `ProfileEvent_USearchSearchCount` UInt64 COMMENT 'Number of search operations performed in usearch indexes.', `ProfileEvent_USearchSearchVisitedMembers` UInt64 COMMENT 'Number of nodes visited when searching in usearch indexes.', `ProfileEvent_USearchSearchComputedDistances` UInt64 COMMENT 'Number of times distance was computed when searching usearch indexes.', `ProfileEvent_RWLockAcquiredReadLocks` UInt64 COMMENT 'Number of times a read lock was acquired (in a heavy RWLock).', `ProfileEvent_RWLockAcquiredWriteLocks` UInt64 COMMENT 'Number of times a write lock was acquired (in a heavy RWLock).', `ProfileEvent_RWLockReadersWaitMilliseconds` UInt64 COMMENT 'Total time spent waiting for a read lock to be acquired (in a heavy RWLock).', `ProfileEvent_RWLockWritersWaitMilliseconds` UInt64 COMMENT 'Total time spent waiting for a write lock to be acquired (in a heavy RWLock).', `ProfileEvent_DNSError` UInt64 COMMENT 'Total count of errors in DNS resolution', `ProfileEvent_PartsLockHoldMicroseconds` UInt64 COMMENT 'Total time spent holding data parts lock in MergeTree tables', `ProfileEvent_PartsLockWaitMicroseconds` UInt64 COMMENT 'Total time spent waiting for data parts lock in MergeTree tables', `ProfileEvent_RealTimeMicroseconds` UInt64 COMMENT 'Total (wall clock) time spent in processing (queries and other tasks) threads (note that this is a sum).', `ProfileEvent_UserTimeMicroseconds` UInt64 COMMENT 'Total time spent in processing (queries and other tasks) threads executing CPU instructions in user mode. This includes time CPU pipeline was stalled due to main memory access, cache misses, branch mispredictions, hyper-threading, etc.', `ProfileEvent_SystemTimeMicroseconds` UInt64 COMMENT 'Total time spent in processing (queries and other tasks) threads executing CPU instructions in OS kernel mode. This is time spent in syscalls, excluding waiting time during blocking syscalls.', `ProfileEvent_MemoryOvercommitWaitTimeMicroseconds` UInt64 COMMENT 'Total time spent in waiting for memory to be freed in OvercommitTracker.', `ProfileEvent_MemoryAllocatorPurge` UInt64 COMMENT 'Total number of times memory allocator purge was requested', `ProfileEvent_MemoryAllocatorPurgeTimeMicroseconds` UInt64 COMMENT 'Total number of times memory allocator purge was requested', `ProfileEvent_SoftPageFaults` UInt64 COMMENT 'The number of soft page faults in query execution threads. Soft page fault usually means a miss in the memory allocator cache, which requires a new memory mapping from the OS and subsequent allocation of a page of physical memory.', `ProfileEvent_HardPageFaults` UInt64 COMMENT 'The number of hard page faults in query execution threads. High values indicate either that you forgot to turn off swap on your server, or eviction of memory pages of the ClickHouse binary during very high memory pressure, or successful usage of the \'mmap\' read method for the tables data.', `ProfileEvent_OSIOWaitMicroseconds` UInt64 COMMENT 'Total time a thread spent waiting for a result of IO operation, from the OS point of view. This is real IO that doesn\'t include page cache.', `ProfileEvent_OSCPUWaitMicroseconds` UInt64 COMMENT 'Total time a thread was ready for execution but waiting to be scheduled by OS, from the OS point of view.', `ProfileEvent_OSCPUVirtualTimeMicroseconds` UInt64 COMMENT 'CPU time spent seen by OS. Does not include involuntary waits due to virtualization.', `ProfileEvent_OSReadBytes` UInt64 COMMENT 'Number of bytes read from disks or block devices. Doesn\'t include bytes read from page cache. May include excessive data due to block size, readahead, etc.', `ProfileEvent_OSWriteBytes` UInt64 COMMENT 'Number of bytes written to disks or block devices. Doesn\'t include bytes that are in page cache dirty pages. May not include data that was written by OS asynchronously.', `ProfileEvent_OSReadChars` UInt64 COMMENT 'Number of bytes read from filesystem, including page cache.', `ProfileEvent_OSWriteChars` UInt64 COMMENT 'Number of bytes written to filesystem, including page cache.', `ProfileEvent_ParallelReplicasHandleRequestMicroseconds` UInt64 COMMENT 'Time spent processing requests for marks from replicas', `ProfileEvent_ParallelReplicasHandleAnnouncementMicroseconds` UInt64 COMMENT 'Time spent processing replicas announcements', `ProfileEvent_ParallelReplicasAnnouncementMicroseconds` UInt64 COMMENT 'Time spent to send an announcement', `ProfileEvent_ParallelReplicasReadRequestMicroseconds` UInt64 COMMENT 'Time spent for read requests', `ProfileEvent_ParallelReplicasReadAssignedMarks` UInt64 COMMENT 'Sum across all replicas of how many of scheduled marks were assigned by consistent hash', `ProfileEvent_ParallelReplicasReadUnassignedMarks` UInt64 COMMENT 'Sum across all replicas of how many unassigned marks were scheduled', `ProfileEvent_ParallelReplicasReadAssignedForStealingMarks` UInt64 COMMENT 'Sum across all replicas of how many of scheduled marks were assigned for stealing by consistent hash', `ProfileEvent_ParallelReplicasReadMarks` UInt64 COMMENT 'How many marks were read by the given replica', `ProfileEvent_ParallelReplicasStealingByHashMicroseconds` UInt64 COMMENT 'Time spent collecting segments meant for stealing by hash', `ProfileEvent_ParallelReplicasProcessingPartsMicroseconds` UInt64 COMMENT 'Time spent processing data parts', `ProfileEvent_ParallelReplicasStealingLeftoversMicroseconds` UInt64 COMMENT 'Time spent collecting orphaned segments', `ProfileEvent_ParallelReplicasCollectingOwnedSegmentsMicroseconds` UInt64 COMMENT 'Time spent collecting segments meant by hash', `ProfileEvent_ParallelReplicasNumRequests` UInt64 COMMENT 'Number of requests to the initiator.', `ProfileEvent_ParallelReplicasDeniedRequests` UInt64 COMMENT 'Number of completely denied requests to the initiator', `ProfileEvent_CacheWarmerBytesDownloaded` UInt64 COMMENT 'Amount of data fetched into filesystem cache by dedicated background threads.', `ProfileEvent_CacheWarmerDataPartsDownloaded` UInt64 COMMENT 'Number of data parts that were fully fetched by CacheWarmer.', `ProfileEvent_IgnoredColdParts` UInt64 COMMENT 'See setting ignore_cold_parts_seconds. Number of times read queries ignored very new parts that weren\'t pulled into cache by CacheWarmer yet.', `ProfileEvent_PreferredWarmedUnmergedParts` UInt64 COMMENT 'See setting prefer_warmed_unmerged_parts_seconds. Number of times read queries used outdated pre-merge parts that are in cache instead of merged part that wasn\'t pulled into cache by CacheWarmer yet.', `ProfileEvent_PerfCPUCycles` UInt64 COMMENT 'Total cycles. Be wary of what happens during CPU frequency scaling.', `ProfileEvent_PerfInstructions` UInt64 COMMENT 'Retired instructions. Be careful, these can be affected by various issues, most notably hardware interrupt counts.', `ProfileEvent_PerfCacheReferences` UInt64 COMMENT 'Cache accesses. Usually, this indicates Last Level Cache accesses, but this may vary depending on your CPU. This may include prefetches and coherency messages; again this depends on the design of your CPU.', `ProfileEvent_PerfCacheMisses` UInt64 COMMENT 'Cache misses. Usually this indicates Last Level Cache misses; this is intended to be used in conjunction with the PERFCOUNTHWCACHEREFERENCES event to calculate cache miss rates.', `ProfileEvent_PerfBranchInstructions` UInt64 COMMENT 'Retired branch instructions. Prior to Linux 2.6.35, this used the wrong event on AMD processors.', `ProfileEvent_PerfBranchMisses` UInt64 COMMENT 'Mispredicted branch instructions.', `ProfileEvent_PerfBusCycles` UInt64 COMMENT 'Bus cycles, which can be different from total cycles.', `ProfileEvent_PerfStalledCyclesFrontend` UInt64 COMMENT 'Stalled cycles during issue.', `ProfileEvent_PerfStalledCyclesBackend` UInt64 COMMENT 'Stalled cycles during retirement.', `ProfileEvent_PerfRefCPUCycles` UInt64 COMMENT 'Total cycles; not affected by CPU frequency scaling.', `ProfileEvent_PerfCPUClock` UInt64 COMMENT 'The CPU clock, a high-resolution per-CPU timer', `ProfileEvent_PerfTaskClock` UInt64 COMMENT 'A clock count specific to the task that is running', `ProfileEvent_PerfContextSwitches` UInt64 COMMENT 'Number of context switches', `ProfileEvent_PerfCPUMigrations` UInt64 COMMENT 'Number of times the process has migrated to a new CPU', `ProfileEvent_PerfAlignmentFaults` UInt64 COMMENT 'Number of alignment faults. These happen when unaligned memory accesses happen; the kernel can handle these but it reduces performance. This happens only on some architectures (never on x86).', `ProfileEvent_PerfEmulationFaults` UInt64 COMMENT 'Number of emulation faults. The kernel sometimes traps on unimplemented instructions and emulates them for user space. This can negatively impact performance.', `ProfileEvent_PerfMinEnabledTime` UInt64 COMMENT 'For all events, minimum time that an event was enabled. Used to track event multiplexing influence', `ProfileEvent_PerfMinEnabledRunningTime` UInt64 COMMENT 'Running time for event with minimum enabled time. Used to track the amount of event multiplexing', `ProfileEvent_PerfDataTLBReferences` UInt64 COMMENT 'Data TLB references', `ProfileEvent_PerfDataTLBMisses` UInt64 COMMENT 'Data TLB misses', `ProfileEvent_PerfInstructionTLBReferences` UInt64 COMMENT 'Instruction TLB references', `ProfileEvent_PerfInstructionTLBMisses` UInt64 COMMENT 'Instruction TLB misses', `ProfileEvent_PerfLocalMemoryReferences` UInt64 COMMENT 'Local NUMA node memory reads', `ProfileEvent_PerfLocalMemoryMisses` UInt64 COMMENT 'Local NUMA node memory read misses', `ProfileEvent_CannotWriteToWriteBufferDiscard` UInt64 COMMENT 'Number of stack traces dropped by query profiler or signal handler because pipe is full or cannot write to pipe.', `ProfileEvent_QueryProfilerSignalOverruns` UInt64 COMMENT 'Number of times we drop processing of a query profiler signal due to overrun plus the number of signals that OS has not delivered due to overrun.', `ProfileEvent_QueryProfilerConcurrencyOverruns` UInt64 COMMENT 'Number of times we drop processing of a query profiler signal due to too many concurrent query profilers in other threads, which may indicate overload.', `ProfileEvent_QueryProfilerRuns` UInt64 COMMENT 'Number of times QueryProfiler had been run.', `ProfileEvent_QueryProfilerErrors` UInt64 COMMENT 'Invalid memory accesses during asynchronous stack unwinding.', `ProfileEvent_CreatedLogEntryForMerge` UInt64 COMMENT 'Successfully created log entry to merge parts in ReplicatedMergeTree.', `ProfileEvent_NotCreatedLogEntryForMerge` UInt64 COMMENT 'Log entry to merge parts in ReplicatedMergeTree is not created due to concurrent log update by another replica.', `ProfileEvent_CreatedLogEntryForMutation` UInt64 COMMENT 'Successfully created log entry to mutate parts in ReplicatedMergeTree.', `ProfileEvent_NotCreatedLogEntryForMutation` UInt64 COMMENT 'Log entry to mutate parts in ReplicatedMergeTree is not created due to concurrent log update by another replica.', `ProfileEvent_S3ReadMicroseconds` UInt64 COMMENT 'Time of GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsCount` UInt64 COMMENT 'Number of GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsRedirects` UInt64 COMMENT 'Number of redirects in GET and HEAD requests to S3 storage.', `ProfileEvent_S3WriteMicroseconds` UInt64 COMMENT 'Time of POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsCount` UInt64 COMMENT 'Number of POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsRedirects` UInt64 COMMENT 'Number of redirects in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_DiskS3ReadMicroseconds` UInt64 COMMENT 'Time of GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsCount` UInt64 COMMENT 'Number of GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsRedirects` UInt64 COMMENT 'Number of redirects in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteMicroseconds` UInt64 COMMENT 'Time of POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsCount` UInt64 COMMENT 'Number of POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsRedirects` UInt64 COMMENT 'Number of redirects in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_S3DeleteObjects` UInt64 COMMENT 'Number of S3 API DeleteObject(s) calls.', `ProfileEvent_S3CopyObject` UInt64 COMMENT 'Number of S3 API CopyObject calls.', `ProfileEvent_S3ListObjects` UInt64 COMMENT 'Number of S3 API ListObjects calls.', `ProfileEvent_S3HeadObject` UInt64 COMMENT 'Number of S3 API HeadObject calls.', `ProfileEvent_S3GetObjectAttributes` UInt64 COMMENT 'Number of S3 API GetObjectAttributes calls.', `ProfileEvent_S3CreateMultipartUpload` UInt64 COMMENT 'Number of S3 API CreateMultipartUpload calls.', `ProfileEvent_S3UploadPartCopy` UInt64 COMMENT 'Number of S3 API UploadPartCopy calls.', `ProfileEvent_S3UploadPart` UInt64 COMMENT 'Number of S3 API UploadPart calls.', `ProfileEvent_S3AbortMultipartUpload` UInt64 COMMENT 'Number of S3 API AbortMultipartUpload calls.', `ProfileEvent_S3CompleteMultipartUpload` UInt64 COMMENT 'Number of S3 API CompleteMultipartUpload calls.', `ProfileEvent_S3PutObject` UInt64 COMMENT 'Number of S3 API PutObject calls.', `ProfileEvent_S3GetObject` UInt64 COMMENT 'Number of S3 API GetObject calls.', `ProfileEvent_DiskS3DeleteObjects` UInt64 COMMENT 'Number of DiskS3 API DeleteObject(s) calls.', `ProfileEvent_DiskS3CopyObject` UInt64 COMMENT 'Number of DiskS3 API CopyObject calls.', `ProfileEvent_DiskS3ListObjects` UInt64 COMMENT 'Number of DiskS3 API ListObjects calls.', `ProfileEvent_DiskS3HeadObject` UInt64 COMMENT 'Number of DiskS3 API HeadObject calls.', `ProfileEvent_DiskS3GetObjectAttributes` UInt64 COMMENT 'Number of DiskS3 API GetObjectAttributes calls.', `ProfileEvent_DiskS3CreateMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API CreateMultipartUpload calls.', `ProfileEvent_DiskS3UploadPartCopy` UInt64 COMMENT 'Number of DiskS3 API UploadPartCopy calls.', `ProfileEvent_DiskS3UploadPart` UInt64 COMMENT 'Number of DiskS3 API UploadPart calls.', `ProfileEvent_DiskS3AbortMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API AbortMultipartUpload calls.', `ProfileEvent_DiskS3CompleteMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API CompleteMultipartUpload calls.', `ProfileEvent_DiskS3PutObject` UInt64 COMMENT 'Number of DiskS3 API PutObject calls.', `ProfileEvent_DiskS3GetObject` UInt64 COMMENT 'Number of DiskS3 API GetObject calls.', `ProfileEvent_DiskPlainRewritableAzureDirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for AzureObjectStorage.', `ProfileEvent_DiskPlainRewritableAzureDirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for AzureObjectStorage.', `ProfileEvent_DiskPlainRewritableLocalDirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for LocalObjectStorage.', `ProfileEvent_DiskPlainRewritableLocalDirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for LocalObjectStorage.', `ProfileEvent_DiskPlainRewritableS3DirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for S3ObjectStorage.', `ProfileEvent_DiskPlainRewritableS3DirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for S3ObjectStorage.', `ProfileEvent_S3Clients` UInt64 COMMENT 'Number of created S3 clients.', `ProfileEvent_TinyS3Clients` UInt64 COMMENT 'Number of S3 clients copies which reuse an existing auth provider from another client.', `ProfileEvent_EngineFileLikeReadFiles` UInt64 COMMENT 'Number of files read in table engines working with files (like File/S3/URL/HDFS).', `ProfileEvent_ReadBufferFromS3Microseconds` UInt64 COMMENT 'Time spent on reading from S3.', `ProfileEvent_ReadBufferFromS3InitMicroseconds` UInt64 COMMENT 'Time spent initializing connection to S3.', `ProfileEvent_ReadBufferFromS3Bytes` UInt64 COMMENT 'Bytes read from S3.', `ProfileEvent_ReadBufferFromS3RequestsErrors` UInt64 COMMENT 'Number of exceptions while reading from S3.', `ProfileEvent_WriteBufferFromS3Microseconds` UInt64 COMMENT 'Time spent on writing to S3.', `ProfileEvent_WriteBufferFromS3Bytes` UInt64 COMMENT 'Bytes written to S3.', `ProfileEvent_WriteBufferFromS3RequestsErrors` UInt64 COMMENT 'Number of exceptions while writing to S3.', `ProfileEvent_WriteBufferFromS3WaitInflightLimitMicroseconds` UInt64 COMMENT 'Time spent on waiting while some of the current requests are done when its number reached the limit defined by s3_max_inflight_parts_for_one_file.', `ProfileEvent_QueryMemoryLimitExceeded` UInt64 COMMENT 'Number of times when memory limit exceeded for query.', `ProfileEvent_AzureGetObject` UInt64 COMMENT 'Number of Azure API GetObject calls.', `ProfileEvent_AzureUpload` UInt64 COMMENT 'Number of Azure blob storage API Upload calls', `ProfileEvent_AzureStageBlock` UInt64 COMMENT 'Number of Azure blob storage API StageBlock calls', `ProfileEvent_AzureCommitBlockList` UInt64 COMMENT 'Number of Azure blob storage API CommitBlockList calls', `ProfileEvent_AzureCopyObject` UInt64 COMMENT 'Number of Azure blob storage API CopyObject calls', `ProfileEvent_AzureDeleteObjects` UInt64 COMMENT 'Number of Azure blob storage API DeleteObject(s) calls.', `ProfileEvent_AzureListObjects` UInt64 COMMENT 'Number of Azure blob storage API ListObjects calls.', `ProfileEvent_AzureGetProperties` UInt64 COMMENT 'Number of Azure blob storage API GetProperties calls.', `ProfileEvent_AzureCreateContainer` UInt64 COMMENT 'Number of Azure blob storage API CreateContainer calls.', `ProfileEvent_DiskAzureGetObject` UInt64 COMMENT 'Number of Disk Azure API GetObject calls.', `ProfileEvent_DiskAzureUpload` UInt64 COMMENT 'Number of Disk Azure blob storage API Upload calls', `ProfileEvent_DiskAzureStageBlock` UInt64 COMMENT 'Number of Disk Azure blob storage API StageBlock calls', `ProfileEvent_DiskAzureCommitBlockList` UInt64 COMMENT 'Number of Disk Azure blob storage API CommitBlockList calls', `ProfileEvent_DiskAzureCopyObject` UInt64 COMMENT 'Number of Disk Azure blob storage API CopyObject calls', `ProfileEvent_DiskAzureListObjects` UInt64 COMMENT 'Number of Disk Azure blob storage API ListObjects calls.', `ProfileEvent_DiskAzureDeleteObjects` UInt64 COMMENT 'Number of Azure blob storage API DeleteObject(s) calls.', `ProfileEvent_DiskAzureGetProperties` UInt64 COMMENT 'Number of Disk Azure blob storage API GetProperties calls.', `ProfileEvent_DiskAzureCreateContainer` UInt64 COMMENT 'Number of Disk Azure blob storage API CreateContainer calls.', `ProfileEvent_ReadBufferFromAzureMicroseconds` UInt64 COMMENT 'Time spent on reading from Azure.', `ProfileEvent_ReadBufferFromAzureInitMicroseconds` UInt64 COMMENT 'Time spent initializing connection to Azure.', `ProfileEvent_ReadBufferFromAzureBytes` UInt64 COMMENT 'Bytes read from Azure.', `ProfileEvent_ReadBufferFromAzureRequestsErrors` UInt64 COMMENT 'Number of exceptions while reading from Azure', `ProfileEvent_CachedReadBufferReadFromCacheHits` UInt64 COMMENT 'Number of times the read from filesystem cache hit the cache.', `ProfileEvent_CachedReadBufferReadFromCacheMisses` UInt64 COMMENT 'Number of times the read from filesystem cache miss the cache.', `ProfileEvent_CachedReadBufferReadFromSourceMicroseconds` UInt64 COMMENT 'Time reading from filesystem cache source (from remote filesystem, etc)', `ProfileEvent_CachedReadBufferReadFromCacheMicroseconds` UInt64 COMMENT 'Time reading from filesystem cache', `ProfileEvent_CachedReadBufferReadFromSourceBytes` UInt64 COMMENT 'Bytes read from filesystem cache source (from remote fs, etc)', `ProfileEvent_CachedReadBufferReadFromCacheBytes` UInt64 COMMENT 'Bytes read from filesystem cache', `ProfileEvent_CachedReadBufferPredownloadedBytes` UInt64 COMMENT 'Bytes read from filesystem cache source. Cache segments are read from left to right as a whole, it might be that we need to predownload some part of the segment irrelevant for the current task just to get to the needed data', `ProfileEvent_CachedReadBufferCacheWriteBytes` UInt64 COMMENT 'Bytes written from source (remote fs, etc) to filesystem cache', `ProfileEvent_CachedReadBufferCacheWriteMicroseconds` UInt64 COMMENT 'Time spent writing data into filesystem cache', `ProfileEvent_CachedReadBufferCreateBufferMicroseconds` UInt64 COMMENT 'Prepare buffer time', `ProfileEvent_CachedWriteBufferCacheWriteBytes` UInt64 COMMENT 'Bytes written from source (remote fs, etc) to filesystem cache', `ProfileEvent_CachedWriteBufferCacheWriteMicroseconds` UInt64 COMMENT 'Time spent writing data into filesystem cache', `ProfileEvent_FilesystemCacheLoadMetadataMicroseconds` UInt64 COMMENT 'Time spent loading filesystem cache metadata', `ProfileEvent_FilesystemCacheEvictedBytes` UInt64 COMMENT 'Number of bytes evicted from filesystem cache', `ProfileEvent_FilesystemCacheEvictedFileSegments` UInt64 COMMENT 'Number of file segments evicted from filesystem cache', `ProfileEvent_FilesystemCacheBackgroundDownloadQueuePush` UInt64 COMMENT 'Number of file segments sent for background download in filesystem cache', `ProfileEvent_FilesystemCacheEvictionSkippedFileSegments` UInt64 COMMENT 'Number of file segments skipped for eviction because of being in unreleasable state', `ProfileEvent_FilesystemCacheEvictionSkippedEvictingFileSegments` UInt64 COMMENT 'Number of file segments skipped for eviction because of being in evicting state', `ProfileEvent_FilesystemCacheEvictionTries` UInt64 COMMENT 'Number of filesystem cache eviction attempts', `ProfileEvent_FilesystemCacheLockKeyMicroseconds` UInt64 COMMENT 'Lock cache key time', `ProfileEvent_FilesystemCacheLockMetadataMicroseconds` UInt64 COMMENT 'Lock filesystem cache metadata time', `ProfileEvent_FilesystemCacheLockCacheMicroseconds` UInt64 COMMENT 'Lock filesystem cache time', `ProfileEvent_FilesystemCacheReserveMicroseconds` UInt64 COMMENT 'Filesystem cache space reservation time', `ProfileEvent_FilesystemCacheEvictMicroseconds` UInt64 COMMENT 'Filesystem cache eviction time', `ProfileEvent_FilesystemCacheGetOrSetMicroseconds` UInt64 COMMENT 'Filesystem cache getOrSet() time', `ProfileEvent_FilesystemCacheGetMicroseconds` UInt64 COMMENT 'Filesystem cache get() time', `ProfileEvent_FileSegmentWaitMicroseconds` UInt64 COMMENT 'Wait on DOWNLOADING state', `ProfileEvent_FileSegmentCompleteMicroseconds` UInt64 COMMENT 'Duration of FileSegment::complete() in filesystem cache', `ProfileEvent_FileSegmentLockMicroseconds` UInt64 COMMENT 'Lock file segment time', `ProfileEvent_FileSegmentWriteMicroseconds` UInt64 COMMENT 'File segment write() time', `ProfileEvent_FileSegmentUseMicroseconds` UInt64 COMMENT 'File segment use() time', `ProfileEvent_FileSegmentRemoveMicroseconds` UInt64 COMMENT 'File segment remove() time', `ProfileEvent_FileSegmentHolderCompleteMicroseconds` UInt64 COMMENT 'File segments holder complete() time', `ProfileEvent_FileSegmentFailToIncreasePriority` UInt64 COMMENT 'Number of times the priority was not increased due to a high contention on the cache lock', `ProfileEvent_FilesystemCacheFailToReserveSpaceBecauseOfLockContention` UInt64 COMMENT 'Number of times space reservation was skipped due to a high contention on the cache lock', `ProfileEvent_FilesystemCacheFailToReserveSpaceBecauseOfCacheResize` UInt64 COMMENT 'Number of times space reservation was skipped due to the cache is being resized', `ProfileEvent_FilesystemCacheHoldFileSegments` UInt64 COMMENT 'Filesystem cache file segments count, which were hold', `ProfileEvent_FilesystemCacheUnusedHoldFileSegments` UInt64 COMMENT 'Filesystem cache file segments count, which were hold, but not used (because of seek or LIMIT n, etc)', `ProfileEvent_FilesystemCacheFreeSpaceKeepingThreadRun` UInt64 COMMENT 'Number of times background thread executed free space keeping job', `ProfileEvent_FilesystemCacheFreeSpaceKeepingThreadWorkMilliseconds` UInt64 COMMENT 'Time for which background thread executed free space keeping job', `ProfileEvent_RemoteFSSeeks` UInt64 COMMENT 'Total number of seeks for async buffer', `ProfileEvent_RemoteFSPrefetches` UInt64 COMMENT 'Number of prefetches made with asynchronous reading from remote filesystem', `ProfileEvent_RemoteFSCancelledPrefetches` UInt64 COMMENT 'Number of cancelled prefecthes (because of seek)', `ProfileEvent_RemoteFSUnusedPrefetches` UInt64 COMMENT 'Number of prefetches pending at buffer destruction', `ProfileEvent_RemoteFSPrefetchedReads` UInt64 COMMENT 'Number of reads from prefecthed buffer', `ProfileEvent_RemoteFSPrefetchedBytes` UInt64 COMMENT 'Number of bytes from prefecthed buffer', `ProfileEvent_RemoteFSUnprefetchedReads` UInt64 COMMENT 'Number of reads from unprefetched buffer', `ProfileEvent_RemoteFSUnprefetchedBytes` UInt64 COMMENT 'Number of bytes from unprefetched buffer', `ProfileEvent_RemoteFSLazySeeks` UInt64 COMMENT 'Number of lazy seeks', `ProfileEvent_RemoteFSSeeksWithReset` UInt64 COMMENT 'Number of seeks which lead to a new connection', `ProfileEvent_RemoteFSBuffers` UInt64 COMMENT 'Number of buffers created for asynchronous reading from remote filesystem', `ProfileEvent_MergeTreePrefetchedReadPoolInit` UInt64 COMMENT 'Time spent preparing tasks in MergeTreePrefetchedReadPool', `ProfileEvent_WaitPrefetchTaskMicroseconds` UInt64 COMMENT 'Time spend waiting for prefetched reader', `ProfileEvent_ThreadpoolReaderTaskMicroseconds` UInt64 COMMENT 'Time spent getting the data in asynchronous reading', `ProfileEvent_ThreadpoolReaderPrepareMicroseconds` UInt64 COMMENT 'Time spent on preparation (e.g. call to reader seek() method)', `ProfileEvent_ThreadpoolReaderReadBytes` UInt64 COMMENT 'Bytes read from a threadpool task in asynchronous reading', `ProfileEvent_ThreadpoolReaderSubmit` UInt64 COMMENT 'Bytes read from a threadpool task in asynchronous reading', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronously` UInt64 COMMENT 'How many times we haven\'t scheduled a task on the thread pool and read synchronously instead', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronouslyBytes` UInt64 COMMENT 'How many bytes were read synchronously', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronouslyMicroseconds` UInt64 COMMENT 'How much time we spent reading synchronously', `ProfileEvent_ThreadpoolReaderSubmitLookupInCacheMicroseconds` UInt64 COMMENT 'How much time we spent checking if content is cached', `ProfileEvent_AsynchronousReaderIgnoredBytes` UInt64 COMMENT 'Number of bytes ignored during asynchronous reading', `ProfileEvent_FileSegmentWaitReadBufferMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend waiting for internal read buffer (includes cache waiting)', `ProfileEvent_FileSegmentReadMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend reading from file', `ProfileEvent_FileSegmentCacheWriteMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend writing data to cache', `ProfileEvent_FileSegmentPredownloadMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spent pre-downloading data to cache (pre-downloading - finishing file segment download (after someone who failed to do that) up to the point current thread was requested to do)', `ProfileEvent_FileSegmentUsedBytes` UInt64 COMMENT 'Metric per file segment. How many bytes were actually used from current file segment', `ProfileEvent_ReadBufferSeekCancelConnection` UInt64 COMMENT 'Number of seeks which lead to new connection (s3, http)', `ProfileEvent_SleepFunctionCalls` UInt64 COMMENT 'Number of times a sleep function (sleep, sleepEachRow) has been called.', `ProfileEvent_SleepFunctionMicroseconds` UInt64 COMMENT 'Time set to sleep in a sleep function (sleep, sleepEachRow).', `ProfileEvent_SleepFunctionElapsedMicroseconds` UInt64 COMMENT 'Time spent sleeping in a sleep function (sleep, sleepEachRow).', `ProfileEvent_ThreadPoolReaderPageCacheHit` UInt64 COMMENT 'Number of times the read inside ThreadPoolReader was done from the page cache.', `ProfileEvent_ThreadPoolReaderPageCacheHitBytes` UInt64 COMMENT 'Number of bytes read inside ThreadPoolReader when it was done from the page cache.', `ProfileEvent_ThreadPoolReaderPageCacheHitElapsedMicroseconds` UInt64 COMMENT 'Time spent reading data from page cache in ThreadPoolReader.', `ProfileEvent_ThreadPoolReaderPageCacheMiss` UInt64 COMMENT 'Number of times the read inside ThreadPoolReader was not done from page cache and was hand off to thread pool.', `ProfileEvent_ThreadPoolReaderPageCacheMissBytes` UInt64 COMMENT 'Number of bytes read inside ThreadPoolReader when read was not done from page cache and was hand off to thread pool.', `ProfileEvent_ThreadPoolReaderPageCacheMissElapsedMicroseconds` UInt64 COMMENT 'Time spent reading data inside the asynchronous job in ThreadPoolReader - when read was not done from the page cache.', `ProfileEvent_AsynchronousReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for asynchronous reads in asynchronous local read.', `ProfileEvent_SynchronousReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for synchronous reads in asynchronous local read.', `ProfileEvent_AsynchronousRemoteReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for asynchronous remote reads.', `ProfileEvent_SynchronousRemoteReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for synchronous remote reads.', `ProfileEvent_ExternalDataSourceLocalCacheReadBytes` UInt64 COMMENT 'Bytes read from local cache buffer in RemoteReadBufferCache', `ProfileEvent_MainConfigLoads` UInt64 COMMENT 'Number of times the main configuration was reloaded.', `ProfileEvent_AggregationPreallocatedElementsInHashTables` UInt64 COMMENT 'How many elements were preallocated in hash tables for aggregation.', `ProfileEvent_AggregationHashTablesInitializedAsTwoLevel` UInt64 COMMENT 'How many hash tables were inited as two-level for aggregation.', `ProfileEvent_AggregationOptimizedEqualRangesOfKeys` UInt64 COMMENT 'For how many blocks optimization of equal ranges of keys was applied', `ProfileEvent_HashJoinPreallocatedElementsInHashTables` UInt64 COMMENT 'How many elements were preallocated in hash tables for hash join.', `ProfileEvent_MetadataFromKeeperCacheHit` UInt64 COMMENT 'Number of times an object storage metadata request was answered from cache without making request to Keeper', `ProfileEvent_MetadataFromKeeperCacheMiss` UInt64 COMMENT 'Number of times an object storage metadata request had to be answered from Keeper', `ProfileEvent_MetadataFromKeeperCacheUpdateMicroseconds` UInt64 COMMENT 'Total time spent in updating the cache including waiting for responses from Keeper', `ProfileEvent_MetadataFromKeeperUpdateCacheOneLevel` UInt64 COMMENT 'Number of times a cache update for one level of directory tree was done', `ProfileEvent_MetadataFromKeeperTransactionCommit` UInt64 COMMENT 'Number of times metadata transaction commit was attempted', `ProfileEvent_MetadataFromKeeperTransactionCommitRetry` UInt64 COMMENT 'Number of times metadata transaction commit was retried', `ProfileEvent_MetadataFromKeeperCleanupTransactionCommit` UInt64 COMMENT 'Number of times metadata transaction commit for deleted objects cleanup was attempted', `ProfileEvent_MetadataFromKeeperCleanupTransactionCommitRetry` UInt64 COMMENT 'Number of times metadata transaction commit for deleted objects cleanup was retried', `ProfileEvent_MetadataFromKeeperOperations` UInt64 COMMENT 'Number of times a request was made to Keeper', `ProfileEvent_MetadataFromKeeperIndividualOperations` UInt64 COMMENT 'Number of paths read or written by single or multi requests to Keeper', `ProfileEvent_MetadataFromKeeperReconnects` UInt64 COMMENT 'Number of times a reconnect to Keeper was done', `ProfileEvent_MetadataFromKeeperBackgroundCleanupObjects` UInt64 COMMENT 'Number of times a old deleted object clean up was performed by background task', `ProfileEvent_MetadataFromKeeperBackgroundCleanupTransactions` UInt64 COMMENT 'Number of times old transaction idempotency token was cleaned up by background task', `ProfileEvent_MetadataFromKeeperBackgroundCleanupErrors` UInt64 COMMENT 'Number of times an error was encountered in background cleanup task', `ProfileEvent_SharedMergeTreeMetadataCacheHintLoadedFromCache` UInt64 COMMENT 'Number of times metadata cache hint was found without going to Keeper', `ProfileEvent_KafkaRebalanceRevocations` UInt64 COMMENT 'Number of partition revocations (the first stage of consumer group rebalance)', `ProfileEvent_KafkaRebalanceAssignments` UInt64 COMMENT 'Number of partition assignments (the final stage of consumer group rebalance)', `ProfileEvent_KafkaRebalanceErrors` UInt64 COMMENT 'Number of failed consumer group rebalances', `ProfileEvent_KafkaMessagesPolled` UInt64 COMMENT 'Number of Kafka messages polled from librdkafka to ClickHouse', `ProfileEvent_KafkaMessagesRead` UInt64 COMMENT 'Number of Kafka messages already processed by ClickHouse', `ProfileEvent_KafkaMessagesFailed` UInt64 COMMENT 'Number of Kafka messages ClickHouse failed to parse', `ProfileEvent_KafkaRowsRead` UInt64 COMMENT 'Number of rows parsed from Kafka messages', `ProfileEvent_KafkaRowsRejected` UInt64 COMMENT 'Number of parsed rows which were later rejected (due to rebalances / errors or similar reasons). Those rows will be consumed again after the rebalance.', `ProfileEvent_KafkaDirectReads` UInt64 COMMENT 'Number of direct selects from Kafka tables since server start', `ProfileEvent_KafkaBackgroundReads` UInt64 COMMENT 'Number of background reads populating materialized views from Kafka since server start', `ProfileEvent_KafkaCommits` UInt64 COMMENT 'Number of successful commits of consumed offsets to Kafka (normally should be the same as KafkaBackgroundReads)', `ProfileEvent_KafkaCommitFailures` UInt64 COMMENT 'Number of failed commits of consumed offsets to Kafka (usually is a sign of some data duplication)', `ProfileEvent_KafkaConsumerErrors` UInt64 COMMENT 'Number of errors reported by librdkafka during polls', `ProfileEvent_KafkaWrites` UInt64 COMMENT 'Number of writes (inserts) to Kafka tables ', `ProfileEvent_KafkaRowsWritten` UInt64 COMMENT 'Number of rows inserted into Kafka tables', `ProfileEvent_KafkaProducerFlushes` UInt64 COMMENT 'Number of explicit flushes to Kafka producer', `ProfileEvent_KafkaMessagesProduced` UInt64 COMMENT 'Number of messages produced to Kafka', `ProfileEvent_KafkaProducerErrors` UInt64 COMMENT 'Number of errors during producing the messages to Kafka', `ProfileEvent_ScalarSubqueriesGlobalCacheHit` UInt64 COMMENT 'Number of times a read from a scalar subquery was done using the global cache', `ProfileEvent_ScalarSubqueriesLocalCacheHit` UInt64 COMMENT 'Number of times a read from a scalar subquery was done using the local cache', `ProfileEvent_ScalarSubqueriesCacheMiss` UInt64 COMMENT 'Number of times a read from a scalar subquery was not cached and had to be calculated completely', `ProfileEvent_SchemaInferenceCacheHits` UInt64 COMMENT 'Number of times the requested source is found in schema cache', `ProfileEvent_SchemaInferenceCacheSchemaHits` UInt64 COMMENT 'Number of times the schema is found in schema cache during schema inference', `ProfileEvent_SchemaInferenceCacheNumRowsHits` UInt64 COMMENT 'Number of times the number of rows is found in schema cache during count from files', `ProfileEvent_SchemaInferenceCacheMisses` UInt64 COMMENT 'Number of times the requested source is not in schema cache', `ProfileEvent_SchemaInferenceCacheSchemaMisses` UInt64 COMMENT 'Number of times the requested source is in cache but the schema is not in cache during schema inference', `ProfileEvent_SchemaInferenceCacheNumRowsMisses` UInt64 COMMENT 'Number of times the requested source is in cache but the number of rows is not in cache while count from files', `ProfileEvent_SchemaInferenceCacheEvictions` UInt64 COMMENT 'Number of times a schema from cache was evicted due to overflow', `ProfileEvent_SchemaInferenceCacheInvalidations` UInt64 COMMENT 'Number of times a schema in cache became invalid due to changes in data', `ProfileEvent_KeeperPacketsSent` UInt64 COMMENT 'Packets sent by keeper server', `ProfileEvent_KeeperPacketsReceived` UInt64 COMMENT 'Packets received by keeper server', `ProfileEvent_KeeperRequestTotal` UInt64 COMMENT 'Total requests number on keeper server', `ProfileEvent_KeeperLatency` UInt64 COMMENT 'Keeper latency', `ProfileEvent_KeeperTotalElapsedMicroseconds` UInt64 COMMENT 'Keeper total latency for a single request', `ProfileEvent_KeeperProcessElapsedMicroseconds` UInt64 COMMENT 'Keeper commit latency for a single request', `ProfileEvent_KeeperPreprocessElapsedMicroseconds` UInt64 COMMENT 'Keeper preprocessing latency for a single reuquest', `ProfileEvent_KeeperStorageLockWaitMicroseconds` UInt64 COMMENT 'Time spent waiting for acquiring Keeper storage lock', `ProfileEvent_KeeperCommitWaitElapsedMicroseconds` UInt64 COMMENT 'Time spent waiting for certain log to be committed', `ProfileEvent_KeeperBatchMaxCount` UInt64 COMMENT 'Number of times the size of batch was limited by the amount', `ProfileEvent_KeeperBatchMaxTotalSize` UInt64 COMMENT 'Number of times the size of batch was limited by the total bytes size', `ProfileEvent_KeeperCommits` UInt64 COMMENT 'Number of successful commits', `ProfileEvent_KeeperCommitsFailed` UInt64 COMMENT 'Number of failed commits', `ProfileEvent_KeeperSnapshotCreations` UInt64 COMMENT 'Number of snapshots creations', `ProfileEvent_KeeperSnapshotCreationsFailed` UInt64 COMMENT 'Number of failed snapshot creations', `ProfileEvent_KeeperSnapshotApplys` UInt64 COMMENT 'Number of snapshot applying', `ProfileEvent_KeeperSnapshotApplysFailed` UInt64 COMMENT 'Number of failed snapshot applying', `ProfileEvent_KeeperReadSnapshot` UInt64 COMMENT 'Number of snapshot read(serialization)', `ProfileEvent_KeeperSaveSnapshot` UInt64 COMMENT 'Number of snapshot save', `ProfileEvent_KeeperCreateRequest` UInt64 COMMENT 'Number of create requests', `ProfileEvent_KeeperRemoveRequest` UInt64 COMMENT 'Number of remove requests', `ProfileEvent_KeeperSetRequest` UInt64 COMMENT 'Number of set requests', `ProfileEvent_KeeperReconfigRequest` UInt64 COMMENT 'Number of reconfig requests', `ProfileEvent_KeeperCheckRequest` UInt64 COMMENT 'Number of check requests', `ProfileEvent_KeeperMultiRequest` UInt64 COMMENT 'Number of multi requests', `ProfileEvent_KeeperMultiReadRequest` UInt64 COMMENT 'Number of multi read requests', `ProfileEvent_KeeperGetRequest` UInt64 COMMENT 'Number of get requests', `ProfileEvent_KeeperListRequest` UInt64 COMMENT 'Number of list requests', `ProfileEvent_KeeperExistsRequest` UInt64 COMMENT 'Number of exists requests', `ProfileEvent_OverflowBreak` UInt64 COMMENT 'Number of times, data processing was cancelled by query complexity limitation with setting \'*_overflow_mode\' = \'break\' and the result is incomplete.', `ProfileEvent_OverflowThrow` UInt64 COMMENT 'Number of times, data processing was cancelled by query complexity limitation with setting \'*_overflow_mode\' = \'throw\' and exception was thrown.', `ProfileEvent_OverflowAny` UInt64 COMMENT 'Number of times approximate GROUP BY was in effect: when aggregation was performed only on top of first \'max_rows_to_group_by\' unique keys and other keys were ignored due to \'group_by_overflow_mode\' = \'any\'.', `ProfileEvent_S3QueueSetFileProcessingMicroseconds` UInt64 COMMENT 'Time spent to set file as processing', `ProfileEvent_S3QueueSetFileProcessedMicroseconds` UInt64 COMMENT 'Time spent to set file as processed', `ProfileEvent_S3QueueSetFileFailedMicroseconds` UInt64 COMMENT 'Time spent to set file as failed', `ProfileEvent_ObjectStorageQueueFailedFiles` UInt64 COMMENT 'Number of files which failed to be processed', `ProfileEvent_ObjectStorageQueueProcessedFiles` UInt64 COMMENT 'Number of files which were processed', `ProfileEvent_ObjectStorageQueueCleanupMaxSetSizeOrTTLMicroseconds` UInt64 COMMENT 'Time spent to set file as failed', `ProfileEvent_ObjectStorageQueuePullMicroseconds` UInt64 COMMENT 'Time spent to read file data', `ProfileEvent_ObjectStorageQueueLockLocalFileStatusesMicroseconds` UInt64 COMMENT 'Time spent to lock local file statuses', `ProfileEvent_ObjectStorageQueueFailedToBatchSetProcessing` UInt64 COMMENT 'Number of times batched set processing request failed', `ProfileEvent_ObjectStorageQueueTrySetProcessingRequests` UInt64 COMMENT 'The number of times we tried to make set processing request', `ProfileEvent_ObjectStorageQueueTrySetProcessingSucceeded` UInt64 COMMENT 'The number of times we successfully set file as processing', `ProfileEvent_ObjectStorageQueueTrySetProcessingFailed` UInt64 COMMENT 'The number of times we unsuccessfully set file as processing', `ProfileEvent_ObjectStorageQueueListedFiles` UInt64 COMMENT 'Number of listed files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueFilteredFiles` UInt64 COMMENT 'Number of filtered files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueReadFiles` UInt64 COMMENT 'Number of read files (not equal to the number of actually inserted files)', `ProfileEvent_ObjectStorageQueueReadRows` UInt64 COMMENT 'Number of read rows (not equal to the number of actually inserted rows)', `ProfileEvent_ObjectStorageQueueReadBytes` UInt64 COMMENT 'Number of read bytes (not equal to the number of actually inserted bytes)', `ProfileEvent_ObjectStorageQueueExceptionsDuringRead` UInt64 COMMENT 'Number of exceptions during read in S3(Azure)Queue', `ProfileEvent_ObjectStorageQueueExceptionsDuringInsert` UInt64 COMMENT 'Number of exceptions during insert in S3(Azure)Queue', `ProfileEvent_ObjectStorageQueueRemovedObjects` UInt64 COMMENT 'Number of objects removed as part of after_processing = delete', `ProfileEvent_ObjectStorageQueueInsertIterations` UInt64 COMMENT 'Number of insert iterations', `ProfileEvent_ObjectStorageQueueCommitRequests` UInt64 COMMENT 'Number of keeper requests to commit files as either failed or processed', `ProfileEvent_ObjectStorageQueueSuccessfulCommits` UInt64 COMMENT 'Number of successful keeper commits', `ProfileEvent_ObjectStorageQueueUnsuccessfulCommits` UInt64 COMMENT 'Number of unsuccessful keeper commits', `ProfileEvent_ObjectStorageQueueCancelledFiles` UInt64 COMMENT 'Number cancelled files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueProcessedRows` UInt64 COMMENT 'Number of processed rows in StorageS3(Azure)Queue', `ProfileEvent_ServerStartupMilliseconds` UInt64 COMMENT 'Time elapsed from starting server to listening to sockets in milliseconds', `ProfileEvent_IOUringSQEsSubmitted` UInt64 COMMENT 'Total number of io_uring SQEs submitted', `ProfileEvent_IOUringSQEsResubmitsAsync` UInt64 COMMENT 'Total number of asynchronous io_uring SQE resubmits performed', `ProfileEvent_IOUringSQEsResubmitsSync` UInt64 COMMENT 'Total number of synchronous io_uring SQE resubmits performed', `ProfileEvent_IOUringCQEsCompleted` UInt64 COMMENT 'Total number of successfully completed io_uring CQEs', `ProfileEvent_IOUringCQEsFailed` UInt64 COMMENT 'Total number of completed io_uring CQEs with failures', `ProfileEvent_BackupsOpenedForRead` UInt64 COMMENT 'Number of backups opened for reading', `ProfileEvent_BackupsOpenedForWrite` UInt64 COMMENT 'Number of backups opened for writing', `ProfileEvent_BackupsOpenedForUnlock` UInt64 COMMENT 'Number of backups opened for unlocking', `ProfileEvent_BackupReadMetadataMicroseconds` UInt64 COMMENT 'Time spent reading backup metadata from .backup file', `ProfileEvent_BackupWriteMetadataMicroseconds` UInt64 COMMENT 'Time spent writing backup metadata to .backup file', `ProfileEvent_BackupEntriesCollectorMicroseconds` UInt64 COMMENT 'Time spent making backup entries', `ProfileEvent_BackupEntriesCollectorForTablesDataMicroseconds` UInt64 COMMENT 'Time spent making backup entries for tables data', `ProfileEvent_BackupEntriesCollectorRunPostTasksMicroseconds` UInt64 COMMENT 'Time spent running post tasks after making backup entries', `ProfileEvent_BackupPreparingFileInfosMicroseconds` UInt64 COMMENT 'Time spent preparing file infos for backup entries', `ProfileEvent_BackupReadLocalFilesToCalculateChecksums` UInt64 COMMENT 'Number of files read locally to calculate checksums for backup entries', `ProfileEvent_BackupReadLocalBytesToCalculateChecksums` UInt64 COMMENT 'Total size of files read locally to calculate checksums for backup entries', `ProfileEvent_BackupReadRemoteFilesToCalculateChecksums` UInt64 COMMENT 'Number of files read from remote disks to calculate checksums for backup entries', `ProfileEvent_BackupReadRemoteBytesToCalculateChecksums` UInt64 COMMENT 'Total size of files read from remote disks to calculate checksums for backup entries', `ProfileEvent_BackupLockFileReads` UInt64 COMMENT 'How many times the \'.lock\' file was read while making backup', `ProfileEvent_RestorePartsSkippedFiles` UInt64 COMMENT 'Number of files skipped while restoring parts', `ProfileEvent_RestorePartsSkippedBytes` UInt64 COMMENT 'Total size of files skipped while restoring parts', `ProfileEvent_ReadTaskRequestsReceived` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the initiator server side.', `ProfileEvent_MergeTreeReadTaskRequestsReceived` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the initiator server side.', `ProfileEvent_ReadTaskRequestsSent` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `ProfileEvent_MergeTreeReadTaskRequestsSent` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergeTreeAllRangesAnnouncementsSent` UInt64 COMMENT 'The number of announcements sent from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_ReadTaskRequestsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `ProfileEvent_MergeTreeReadTaskRequestsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergeTreeAllRangesAnnouncementsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in sending the announcement from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergerMutatorsGetPartsForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to take data parts snapshot to build ranges from them.', `ProfileEvent_MergerMutatorPrepareRangesForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to prepare parts ranges which can be merged according to merge predicate.', `ProfileEvent_MergerMutatorSelectPartsForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to select parts from ranges which can be merged.', `ProfileEvent_MergerMutatorRangesForMergeCount` UInt64 COMMENT 'Amount of candidate ranges for merge', `ProfileEvent_MergerMutatorPartsInRangesForMergeCount` UInt64 COMMENT 'Amount of candidate parts for merge', `ProfileEvent_MergerMutatorSelectRangePartsCount` UInt64 COMMENT 'Amount of parts in selected range for merge', `ProfileEvent_ConnectionPoolIsFullMicroseconds` UInt64 COMMENT 'Total time spent waiting for a slot in connection pool.', `ProfileEvent_AsyncLoaderWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting for async loader jobs.', `ProfileEvent_DistrCacheServerSwitches` UInt64 COMMENT 'Distributed Cache read buffer event. Number of server switches between distributed cache servers in read/write-through cache', `ProfileEvent_DistrCacheReadMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent reading from distributed cache', `ProfileEvent_DistrCacheFallbackReadMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spend reading from fallback buffer instead of distributed cache', `ProfileEvent_DistrCachePrecomputeRangesMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent to precompute read ranges', `ProfileEvent_DistrCacheNextImplMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spend in ReadBufferFromDistributedCache::nextImpl', `ProfileEvent_DistrCacheStartRangeMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent to start a new read range with distributed cache', `ProfileEvent_DistrCacheIgnoredBytesWhileWaitingProfileEvents` UInt64 COMMENT 'Distributed Cache read buffer event. Ignored bytes while waiting for profile events in distributed cache', `ProfileEvent_DistrCacheRangeChange` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we changed read range because of seek/last_position change', `ProfileEvent_DistrCacheGetResponseMicroseconds` UInt64 COMMENT 'Distributed Cache client event. Time spend to wait for response from distributed cache', `ProfileEvent_DistrCacheReadErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors during read', `ProfileEvent_DistrCacheMakeRequestErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors when making a request', `ProfileEvent_DistrCacheReceiveResponseErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors when receiving response a request', `ProfileEvent_DistrCachePackets` UInt64 COMMENT 'Distributed Cache client event. Total number of packets received from distributed cache', `ProfileEvent_DistrCachePacketsBytes` UInt64 COMMENT 'Distributed Cache client event. The number of bytes in Data packets which were not ignored', `ProfileEvent_DistrCacheUnusedPackets` UInt64 COMMENT 'Distributed Cache client event. Number of skipped unused packets from distributed cache', `ProfileEvent_DistrCacheUnusedPacketsBytes` UInt64 COMMENT 'Distributed Cache client event. The number of bytes in Data packets which were ignored', `ProfileEvent_DistrCacheUnusedPacketsBufferAllocations` UInt64 COMMENT 'Distributed Cache client event. The number of extra buffer allocations in case we could not reuse existing buffer', `ProfileEvent_DistrCacheLockRegistryMicroseconds` UInt64 COMMENT 'Distributed Cache registry event. Time spent to take DistributedCacheRegistry lock', `ProfileEvent_DistrCacheRegistryUpdateMicroseconds` UInt64 COMMENT 'Distributed Cache registry event. Time spent updating distributed cache registry', `ProfileEvent_DistrCacheRegistryUpdates` UInt64 COMMENT 'Distributed Cache registry event. Number of distributed cache registry updates', `ProfileEvent_DistrCacheHashRingRebuilds` UInt64 COMMENT 'Distributed Cache registry event. Number of distributed cache hash ring rebuilds', `ProfileEvent_DistrCacheReadBytesFromCache` UInt64 COMMENT 'Distributed Cache read buffer event. Bytes read from distributed cache', `ProfileEvent_DistrCacheReadBytesFromFallbackBuffer` UInt64 COMMENT 'Distributed Cache read buffer event. Bytes read from fallback buffer', `ProfileEvent_DistrCacheRangeResetBackward` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we reset read range because of seek/last_position change', `ProfileEvent_DistrCacheRangeResetForward` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we reset read range because of seek/last_position change', `ProfileEvent_DistrCacheOpenedConnections` UInt64 COMMENT 'Distributed Cache connection event. The number of open connections to distributed cache', `ProfileEvent_DistrCacheReusedConnections` UInt64 COMMENT 'Distributed Cache connection event. The number of reused connections to distributed cache', `ProfileEvent_DistrCacheOpenedConnectionsBypassingPool` UInt64 COMMENT 'Distributed Cache connection event. The number of open connections to distributed cache bypassing pool', `ProfileEvent_DistrCacheConnectMicroseconds` UInt64 COMMENT 'Distributed Cache connection event. The time spent to connect to distributed cache', `ProfileEvent_DistrCacheConnectAttempts` UInt64 COMMENT 'Distributed Cache connection event. The number of connection attempts to distributed cache', `ProfileEvent_DistrCacheGetClientMicroseconds` UInt64 COMMENT 'Distributed Cache connection event. Time spent getting client for distributed cache', `ProfileEvent_DistrCacheServerProcessRequestMicroseconds` UInt64 COMMENT 'Distributed Cache server event. Time spent processing request on DistributedCache server side', `ProfileEvent_DistrCacheServerStartRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of StartRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerContinueRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of ContinueRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerEndRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of EndRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerAckRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of AckRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerNewS3CachedClients` UInt64 COMMENT 'Distributed Cache server event. The number of new cached s3 clients', `ProfileEvent_DistrCacheServerReusedS3CachedClients` UInt64 COMMENT 'Distributed Cache server event. The number of reused cached s3 clients', `ProfileEvent_LogTest` UInt64 COMMENT 'Number of log messages with level Test', `ProfileEvent_LogTrace` UInt64 COMMENT 'Number of log messages with level Trace', `ProfileEvent_LogDebug` UInt64 COMMENT 'Number of log messages with level Debug', `ProfileEvent_LogInfo` UInt64 COMMENT 'Number of log messages with level Info', `ProfileEvent_LogWarning` UInt64 COMMENT 'Number of log messages with level Warning', `ProfileEvent_LogError` UInt64 COMMENT 'Number of log messages with level Error', `ProfileEvent_LogFatal` UInt64 COMMENT 'Number of log messages with level Fatal', `ProfileEvent_LoggerElapsedNanoseconds` UInt64 COMMENT 'Cumulative time spend in logging', `ProfileEvent_InterfaceHTTPSendBytes` UInt64 COMMENT 'Number of bytes sent through HTTP interfaces', `ProfileEvent_InterfaceHTTPReceiveBytes` UInt64 COMMENT 'Number of bytes received through HTTP interfaces', `ProfileEvent_InterfaceNativeSendBytes` UInt64 COMMENT 'Number of bytes sent through native interfaces', `ProfileEvent_InterfaceNativeReceiveBytes` UInt64 COMMENT 'Number of bytes received through native interfaces', `ProfileEvent_InterfacePrometheusSendBytes` UInt64 COMMENT 'Number of bytes sent through Prometheus interfaces', `ProfileEvent_InterfacePrometheusReceiveBytes` UInt64 COMMENT 'Number of bytes received through Prometheus interfaces', `ProfileEvent_InterfaceInterserverSendBytes` UInt64 COMMENT 'Number of bytes sent through interserver interfaces', `ProfileEvent_InterfaceInterserverReceiveBytes` UInt64 COMMENT 'Number of bytes received through interserver interfaces', `ProfileEvent_InterfaceMySQLSendBytes` UInt64 COMMENT 'Number of bytes sent through MySQL interfaces', `ProfileEvent_InterfaceMySQLReceiveBytes` UInt64 COMMENT 'Number of bytes received through MySQL interfaces', `ProfileEvent_InterfacePostgreSQLSendBytes` UInt64 COMMENT 'Number of bytes sent through PostgreSQL interfaces', `ProfileEvent_InterfacePostgreSQLReceiveBytes` UInt64 COMMENT 'Number of bytes received through PostgreSQL interfaces', `ProfileEvent_ParallelReplicasUsedCount` UInt64 COMMENT 'Number of replicas used to execute a query with task-based parallel replicas', `ProfileEvent_ParallelReplicasAvailableCount` UInt64 COMMENT 'Number of replicas available to execute a query with task-based parallel replicas', `ProfileEvent_ParallelReplicasUnavailableCount` UInt64 COMMENT 'Number of replicas which was chosen, but found to be unavailable during query execution with task-based parallel replicas', `ProfileEvent_SharedMergeTreeVirtualPartsUpdates` UInt64 COMMENT 'Virtual parts update count', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesByLeader` UInt64 COMMENT 'Virtual parts updates by leader', `ProfileEvent_SharedMergeTreeVirtualPartsUpdateMicroseconds` UInt64 COMMENT 'Virtual parts update microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromZooKeeper` UInt64 COMMENT 'Virtual parts updates count from ZooKeeper', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromZooKeeperMicroseconds` UInt64 COMMENT 'Virtual parts updates from ZooKeeper microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesPeerNotFound` UInt64 COMMENT 'Virtual updates from peer failed because no one found', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromPeer` UInt64 COMMENT 'Virtual parts updates count from peer', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromPeerMicroseconds` UInt64 COMMENT 'Virtual parts updates from peer microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesForMergesOrStatus` UInt64 COMMENT 'Virtual parts updates from non-default background job', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesLeaderFailedElection` UInt64 COMMENT 'Virtual parts updates leader election failed', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesLeaderSuccessfulElection` UInt64 COMMENT 'Virtual parts updates leader election successful', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentAttempt` UInt64 COMMENT 'How many times we tried to assign merge or mutation', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentFailedWithNothingToDo` UInt64 COMMENT 'How many times we tried to assign merge or mutation and failed because nothing to merge', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentFailedWithConflict` UInt64 COMMENT 'How many times we tried to assign merge or mutation and failed because of conflict in Keeper', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentSuccessful` UInt64 COMMENT 'How many times we tried to assign merge or mutation', `ProfileEvent_SharedMergeTreeMergePartsMovedToOudated` UInt64 COMMENT 'How many parts moved to oudated directory', `ProfileEvent_SharedMergeTreeMergePartsMovedToCondemned` UInt64 COMMENT 'How many parts moved to condemned directory', `ProfileEvent_SharedMergeTreeOutdatedPartsConfirmationRequest` UInt64 COMMENT 'How many ZooKeeper requests were used to config outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsConfirmationInvocations` UInt64 COMMENT 'How many invocations were made to confirm outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsHTTPRequest` UInt64 COMMENT 'How many HTTP requests were send to confirm outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsHTTPResponse` UInt64 COMMENT 'How many HTTP responses were send to confirm outdated parts', `ProfileEvent_SharedMergeTreeCondemnedPartsKillRequest` UInt64 COMMENT 'How many ZooKeeper requests were used to remove condemned parts', `ProfileEvent_SharedMergeTreeCondemnedPartsLockConfict` UInt64 COMMENT 'How many times we failed to acquite lock because of conflict', `ProfileEvent_SharedMergeTreeCondemnedPartsRemoved` UInt64 COMMENT 'How many condemned parts were removed', `ProfileEvent_KeeperLogsEntryReadFromLatestCache` UInt64 COMMENT 'Number of log entries in Keeper being read from latest logs cache', `ProfileEvent_KeeperLogsEntryReadFromCommitCache` UInt64 COMMENT 'Number of log entries in Keeper being read from commit logs cache', `ProfileEvent_KeeperLogsEntryReadFromFile` UInt64 COMMENT 'Number of log entries in Keeper being read directly from the changelog file', `ProfileEvent_KeeperLogsPrefetchedEntries` UInt64 COMMENT 'Number of log entries in Keeper being prefetched from the changelog file', `ProfileEvent_StorageConnectionsCreated` UInt64 COMMENT 'Number of created connections for storages', `ProfileEvent_StorageConnectionsReused` UInt64 COMMENT 'Number of reused connections for storages', `ProfileEvent_StorageConnectionsReset` UInt64 COMMENT 'Number of reset connections for storages', `ProfileEvent_StorageConnectionsPreserved` UInt64 COMMENT 'Number of preserved connections for storages', `ProfileEvent_StorageConnectionsExpired` UInt64 COMMENT 'Number of expired connections for storages', `ProfileEvent_StorageConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a connection for storage is failed', `ProfileEvent_StorageConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating connections for storages', `ProfileEvent_DiskConnectionsCreated` UInt64 COMMENT 'Number of created connections for disk', `ProfileEvent_DiskConnectionsReused` UInt64 COMMENT 'Number of reused connections for disk', `ProfileEvent_DiskConnectionsReset` UInt64 COMMENT 'Number of reset connections for disk', `ProfileEvent_DiskConnectionsPreserved` UInt64 COMMENT 'Number of preserved connections for disk', `ProfileEvent_DiskConnectionsExpired` UInt64 COMMENT 'Number of expired connections for disk', `ProfileEvent_DiskConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a connection for disk is failed', `ProfileEvent_DiskConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating connections for disk', `ProfileEvent_HTTPConnectionsCreated` UInt64 COMMENT 'Number of created http connections', `ProfileEvent_HTTPConnectionsReused` UInt64 COMMENT 'Number of reused http connections', `ProfileEvent_HTTPConnectionsReset` UInt64 COMMENT 'Number of reset http connections', `ProfileEvent_HTTPConnectionsPreserved` UInt64 COMMENT 'Number of preserved http connections', `ProfileEvent_HTTPConnectionsExpired` UInt64 COMMENT 'Number of expired http connections', `ProfileEvent_HTTPConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a http connection failed', `ProfileEvent_HTTPConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating http connections', `ProfileEvent_AddressesDiscovered` UInt64 COMMENT 'Total count of new addresses in dns resolve results for http connections', `ProfileEvent_AddressesExpired` UInt64 COMMENT 'Total count of expired addresses which is no longer presented in dns resolve results for http connections', `ProfileEvent_AddressesMarkedAsFailed` UInt64 COMMENT 'Total count of addresses which has been marked as faulty due to connection errors for http connections', `ProfileEvent_ReadWriteBufferFromHTTPRequestsSent` UInt64 COMMENT 'Number of HTTP requests sent by ReadWriteBufferFromHTTP', `ProfileEvent_ReadWriteBufferFromHTTPBytes` UInt64 COMMENT 'Total size of payload bytes received and sent by ReadWriteBufferFromHTTP. Doesn\'t include HTTP headers.', `ProfileEvent_ConcurrencyControlSlotsGranted` UInt64 COMMENT 'Number of CPU slot granted according to guarantee of 1 thread per query and for queries with setting \'use_concurrency_control\' = 0', `ProfileEvent_ConcurrencyControlSlotsDelayed` UInt64 COMMENT 'Number of CPU slot not granted initially and required to wait for a free CPU slot', `ProfileEvent_ConcurrencyControlSlotsAcquired` UInt64 COMMENT 'Total number of CPU slot acquired', `ProfileEvent_ConcurrencyControlSlotsAcquiredNonCompeting` UInt64 COMMENT 'Total number of noncompeting CPU slot acquired', `ProfileEvent_ConcurrencyControlQueriesDelayed` UInt64 COMMENT 'Total number of CPU slot allocations (queries) that were required to wait for slots to upscale', `ProfileEvent_SharedDatabaseCatalogFailedToApplyState` UInt64 COMMENT 'Number of failures to apply new state in SharedDatabaseCatalog', `ProfileEvent_SharedDatabaseCatalogStateApplicationMicroseconds` UInt64 COMMENT 'Total time spend on application of new state in SharedDatabaseCatalog', `ProfileEvent_GWPAsanAllocateSuccess` UInt64 COMMENT 'Number of successful allocations done by GWPAsan', `ProfileEvent_GWPAsanAllocateFailed` UInt64 COMMENT 'Number of failed allocations done by GWPAsan (i.e. filled pool)', `ProfileEvent_GWPAsanFree` UInt64 COMMENT 'Number of free operations done by GWPAsan', `ProfileEvent_MemoryWorkerRun` UInt64 COMMENT 'Number of runs done by MemoryWorker in background', `ProfileEvent_MemoryWorkerRunElapsedMicroseconds` UInt64 COMMENT 'Total time spent by MemoryWorker for background work', `ProfileEvent_ParquetFetchWaitTimeMicroseconds` UInt64 COMMENT 'Time of waiting fetching parquet data' ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains history of memory and metric values from table system.events for individual queries, periodically flushed to disk.\n\nIt is safe to truncate or drop this table at any time.' Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Executing query SHOW CREATE TABLE system.query_thread_log FORMAT TSVRaw on node_default Executing query SHOW CREATE TABLE system.query_thread_log FORMAT TSVRaw on node_default run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['rm', '/var/lib/clickhouse//metadata/db.sql'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['rm', '/var/lib/clickhouse//metadata/db.sql'] Command:[docker exec rootteststorecleanup-gw1-node1-1 rm /var/lib/clickhouse//metadata/db.sql] Command:[docker exec rootteststorecleanup-gw1-node1-1 rm /var/lib/clickhouse//metadata/db.sql] With storage policy, SHOW CREATE TABLE system.query_thread_log is: CREATE TABLE system.query_thread_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `event_date` Date COMMENT 'The date when the thread has finished execution of the query.', `event_time` DateTime COMMENT 'The date and time when the thread has finished execution of the query.', `event_time_microseconds` DateTime64(6) COMMENT 'The date and time when the thread has finished execution of the query with microseconds precision.', `query_start_time` DateTime COMMENT 'Start time of query execution.', `query_start_time_microseconds` DateTime64(6) COMMENT 'Start time of query execution with microsecond precision.', `query_duration_ms` UInt64 COMMENT 'Duration of query execution.', `read_rows` UInt64 COMMENT 'Number of read rows.', `read_bytes` UInt64 COMMENT 'Number of read bytes.', `written_rows` UInt64 COMMENT 'For INSERT queries, the number of written rows. For other queries, the column value is 0.', `written_bytes` UInt64 COMMENT 'For INSERT queries, the number of written bytes. For other queries, the column value is 0.', `memory_usage` Int64 COMMENT 'The difference between the amount of allocated and freed memory in context of this thread.', `peak_memory_usage` Int64 COMMENT 'The maximum difference between the amount of allocated and freed memory in context of this thread.', `thread_name` LowCardinality(String) COMMENT 'Name of the thread.', `thread_id` UInt64 COMMENT 'Internal thread ID.', `master_thread_id` UInt64 COMMENT 'OS initial ID of initial thread.', `current_database` LowCardinality(String) COMMENT 'Name of the current database.', `query` String COMMENT 'Query string.', `normalized_query_hash` UInt64 COMMENT 'The hash of normalized query - with wiped constanstans, etc.', `is_initial_query` UInt8 COMMENT 'Query type. Possible values: 1 — Query was initiated by the client, 0 — Query was initiated by another query for distributed query execution.', `user` LowCardinality(String) COMMENT 'Name of the user who initiated the current query.', `query_id` String COMMENT 'ID of the query.', `address` IPv6 COMMENT 'IP address that was used to make the query.', `port` UInt16 COMMENT 'The client port that was used to make the query.', `initial_user` LowCardinality(String) COMMENT 'Name of the user who ran the initial query (for distributed query execution).', `initial_query_id` String COMMENT 'ID of the initial query (for distributed query execution).', `initial_address` IPv6 COMMENT 'IP address that the parent query was launched from.', `initial_port` UInt16 COMMENT 'The client port that was used to make the parent query.', `initial_query_start_time` DateTime COMMENT 'Start time of the initial query execution.', `initial_query_start_time_microseconds` DateTime64(6) COMMENT 'Start time of the initial query execution ', `interface` UInt8 COMMENT 'Interface that the query was initiated from. Possible values: 1 — TCP, 2 — HTTP.', `is_secure` UInt8 COMMENT 'The flag which shows whether the connection was secure.', `os_user` LowCardinality(String) COMMENT 'OSs username who runs clickhouse-client.', `client_hostname` LowCardinality(String) COMMENT 'Hostname of the client machine where the clickhouse-client or another TCP client is run.', `client_name` LowCardinality(String) COMMENT 'The clickhouse-client or another TCP client name.', `client_revision` UInt32 COMMENT 'Revision of the clickhouse-client or another TCP client.', `client_version_major` UInt32 COMMENT 'Major version of the clickhouse-client or another TCP client.', `client_version_minor` UInt32 COMMENT 'Minor version of the clickhouse-client or another TCP client.', `client_version_patch` UInt32 COMMENT 'Patch component of the clickhouse-client or another TCP client version.', `script_query_number` UInt32 COMMENT 'A sequential query number in a multi-query script.', `script_line_number` UInt32 COMMENT 'A line number in a multi-query script where the current query starts.', `http_method` UInt8 COMMENT 'HTTP method that initiated the query. Possible values: 0 — The query was launched from the TCP interface, 1 — GET method was used., 2 — POST method was used.', `http_user_agent` LowCardinality(String) COMMENT 'The UserAgent header passed in the HTTP request.', `http_referer` String COMMENT 'HTTP header `Referer` passed in the HTTP query (contains an absolute or partial address of the page making the query).', `forwarded_for` String COMMENT 'HTTP header `X-Forwarded-For` passed in the HTTP query.', `quota_key` String COMMENT 'The \'quota key\' specified in the quotas setting.', `distributed_depth` UInt64 COMMENT 'How many times a query was forwarded between servers.', `revision` UInt32 COMMENT 'ClickHouse revision.', `ProfileEvents` Map(LowCardinality(String), UInt64) COMMENT 'ProfileEvents that measure different metrics for this thread. The description of them could be found in the table system.events.', `ProfileEvents.Names` Array(LowCardinality(String)) ALIAS mapKeys(ProfileEvents), `ProfileEvents.Values` Array(UInt64) ALIAS mapValues(ProfileEvents) ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains information about threads that execute queries, for example, thread name, thread start time, duration of query processing.\n\nIt is safe to truncate or drop this table at any time.' With storage policy, SHOW CREATE TABLE system.query_thread_log is: CREATE TABLE system.query_thread_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `event_date` Date COMMENT 'The date when the thread has finished execution of the query.', `event_time` DateTime COMMENT 'The date and time when the thread has finished execution of the query.', `event_time_microseconds` DateTime64(6) COMMENT 'The date and time when the thread has finished execution of the query with microseconds precision.', `query_start_time` DateTime COMMENT 'Start time of query execution.', `query_start_time_microseconds` DateTime64(6) COMMENT 'Start time of query execution with microsecond precision.', `query_duration_ms` UInt64 COMMENT 'Duration of query execution.', `read_rows` UInt64 COMMENT 'Number of read rows.', `read_bytes` UInt64 COMMENT 'Number of read bytes.', `written_rows` UInt64 COMMENT 'For INSERT queries, the number of written rows. For other queries, the column value is 0.', `written_bytes` UInt64 COMMENT 'For INSERT queries, the number of written bytes. For other queries, the column value is 0.', `memory_usage` Int64 COMMENT 'The difference between the amount of allocated and freed memory in context of this thread.', `peak_memory_usage` Int64 COMMENT 'The maximum difference between the amount of allocated and freed memory in context of this thread.', `thread_name` LowCardinality(String) COMMENT 'Name of the thread.', `thread_id` UInt64 COMMENT 'Internal thread ID.', `master_thread_id` UInt64 COMMENT 'OS initial ID of initial thread.', `current_database` LowCardinality(String) COMMENT 'Name of the current database.', `query` String COMMENT 'Query string.', `normalized_query_hash` UInt64 COMMENT 'The hash of normalized query - with wiped constanstans, etc.', `is_initial_query` UInt8 COMMENT 'Query type. Possible values: 1 — Query was initiated by the client, 0 — Query was initiated by another query for distributed query execution.', `user` LowCardinality(String) COMMENT 'Name of the user who initiated the current query.', `query_id` String COMMENT 'ID of the query.', `address` IPv6 COMMENT 'IP address that was used to make the query.', `port` UInt16 COMMENT 'The client port that was used to make the query.', `initial_user` LowCardinality(String) COMMENT 'Name of the user who ran the initial query (for distributed query execution).', `initial_query_id` String COMMENT 'ID of the initial query (for distributed query execution).', `initial_address` IPv6 COMMENT 'IP address that the parent query was launched from.', `initial_port` UInt16 COMMENT 'The client port that was used to make the parent query.', `initial_query_start_time` DateTime COMMENT 'Start time of the initial query execution.', `initial_query_start_time_microseconds` DateTime64(6) COMMENT 'Start time of the initial query execution ', `interface` UInt8 COMMENT 'Interface that the query was initiated from. Possible values: 1 — TCP, 2 — HTTP.', `is_secure` UInt8 COMMENT 'The flag which shows whether the connection was secure.', `os_user` LowCardinality(String) COMMENT 'OSs username who runs clickhouse-client.', `client_hostname` LowCardinality(String) COMMENT 'Hostname of the client machine where the clickhouse-client or another TCP client is run.', `client_name` LowCardinality(String) COMMENT 'The clickhouse-client or another TCP client name.', `client_revision` UInt32 COMMENT 'Revision of the clickhouse-client or another TCP client.', `client_version_major` UInt32 COMMENT 'Major version of the clickhouse-client or another TCP client.', `client_version_minor` UInt32 COMMENT 'Minor version of the clickhouse-client or another TCP client.', `client_version_patch` UInt32 COMMENT 'Patch component of the clickhouse-client or another TCP client version.', `script_query_number` UInt32 COMMENT 'A sequential query number in a multi-query script.', `script_line_number` UInt32 COMMENT 'A line number in a multi-query script where the current query starts.', `http_method` UInt8 COMMENT 'HTTP method that initiated the query. Possible values: 0 — The query was launched from the TCP interface, 1 — GET method was used., 2 — POST method was used.', `http_user_agent` LowCardinality(String) COMMENT 'The UserAgent header passed in the HTTP request.', `http_referer` String COMMENT 'HTTP header `Referer` passed in the HTTP query (contains an absolute or partial address of the page making the query).', `forwarded_for` String COMMENT 'HTTP header `X-Forwarded-For` passed in the HTTP query.', `quota_key` String COMMENT 'The \'quota key\' specified in the quotas setting.', `distributed_depth` UInt64 COMMENT 'How many times a query was forwarded between servers.', `revision` UInt32 COMMENT 'ClickHouse revision.', `ProfileEvents` Map(LowCardinality(String), UInt64) COMMENT 'ProfileEvents that measure different metrics for this thread. The description of them could be found in the table system.events.', `ProfileEvents.Names` Array(LowCardinality(String)) ALIAS mapKeys(ProfileEvents), `ProfileEvents.Values` Array(UInt64) ALIAS mapValues(ProfileEvents) ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains information about threads that execute queries, for example, thread name, thread start time, duration of query processing.\n\nIt is safe to truncate or drop this table at any time.' Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/kek'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/kek'] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/kek] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/kek] Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '0' on node1 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '0' on node1 Executing query SHOW CREATE TABLE system.part_log FORMAT TSVRaw on node_default Executing query SHOW CREATE TABLE system.part_log FORMAT TSVRaw on node_default run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['touch', '/var/lib/clickhouse//store/12'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['touch', '/var/lib/clickhouse//store/12'] Command:[docker exec rootteststorecleanup-gw1-node1-1 touch /var/lib/clickhouse//store/12] Command:[docker exec rootteststorecleanup-gw1-node1-1 touch /var/lib/clickhouse//store/12] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/456'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/456'] Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/456] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/456] With storage policy, SHOW CREATE TABLE system.part_log is: CREATE TABLE system.part_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `query_id` String COMMENT 'Identifier of the INSERT query that created this data part.', `event_type` Enum8('NewPart' = 1, 'MergeParts' = 2, 'DownloadPart' = 3, 'RemovePart' = 4, 'MutatePart' = 5, 'MovePart' = 6, 'MergePartsStart' = 7, 'MutatePartStart' = 8) COMMENT 'Type of the event that occurred with the data part. Can have one of the following values: NewPart — Inserting of a new data part, MergePartsStart — Merging of data parts has started, MergeParts — Merging of data parts has finished, DownloadPart — Downloading a data part, RemovePart — Removing or detaching a data part using [DETACH PARTITION](../../sql-reference/statements/alter/partition.md#alter_detach-partition).MutatePartStart — Mutating of a data part has started, MutatePart — Mutating of a data part has finished, MovePart — Moving the data part from the one disk to another one.', `merge_reason` Enum8('NotAMerge' = 1, 'RegularMerge' = 2, 'TTLDeleteMerge' = 3, 'TTLRecompressMerge' = 4) COMMENT 'The reason for the event with type MERGE_PARTS. Can have one of the following values: NotAMerge — The current event has the type other than MERGE_PARTS, RegularMerge — Some regular merge, TTLDeleteMerge — Cleaning up expired data. TTLRecompressMerge — Recompressing data part with the. ', `merge_algorithm` Enum8('Undecided' = 0, 'Vertical' = 1, 'Horizontal' = 2) COMMENT 'Merge algorithm for the event with type MERGE_PARTS. Can have one of the following values: Undecided, Horizontal, Vertical', `event_date` Date COMMENT 'Event date.', `event_time` DateTime COMMENT 'Event time.', `event_time_microseconds` DateTime64(6) COMMENT 'Event time with microseconds precision.', `duration_ms` UInt64 COMMENT 'Duration of this operation.', `database` String COMMENT 'Name of the database the data part is in.', `table` String COMMENT 'Name of the table the data part is in.', `table_uuid` UUID COMMENT 'UUID of the table the data part belongs to.', `part_name` String COMMENT 'Name of the data part.', `partition_id` String COMMENT 'ID of the partition that the data part was inserted to. The column takes the `all` value if the partitioning is by `tuple()`.', `partition` String COMMENT 'The partition name.', `part_type` String COMMENT 'The type of the part. Possible values: Wide and Compact.', `disk_name` String COMMENT 'The disk name data part lies on.', `path_on_disk` String COMMENT 'Absolute path to the folder with data part files.', `rows` UInt64 COMMENT 'The number of rows in the data part.', `size_in_bytes` UInt64 COMMENT 'Size of the data part on disk in bytes.', `merged_from` Array(String) COMMENT 'An array of the source parts names which the current part was made up from.', `bytes_uncompressed` UInt64 COMMENT 'Uncompressed size of the resulting part in bytes.', `read_rows` UInt64 COMMENT 'The number of rows was read during the merge.', `read_bytes` UInt64 COMMENT 'The number of bytes was read during the merge.', `peak_memory_usage` UInt64 COMMENT 'The maximum amount of used during merge RAM', `error` UInt16 COMMENT 'The error code of the occurred exception.', `exception` String COMMENT 'Text message of the occurred error.', `ProfileEvents` Map(LowCardinality(String), UInt64) COMMENT 'All the profile events captured during this operation.', `ProfileEvents.Names` Array(String) ALIAS mapKeys(ProfileEvents), `ProfileEvents.Values` Array(UInt64) ALIAS mapValues(ProfileEvents), `name` String ALIAS part_name ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'This table contains information about events that occurred with data parts in the MergeTree family tables, such as adding or merging data.\n\nIt is safe to truncate or drop this table at any time.' With storage policy, SHOW CREATE TABLE system.part_log is: CREATE TABLE system.part_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `query_id` String COMMENT 'Identifier of the INSERT query that created this data part.', `event_type` Enum8('NewPart' = 1, 'MergeParts' = 2, 'DownloadPart' = 3, 'RemovePart' = 4, 'MutatePart' = 5, 'MovePart' = 6, 'MergePartsStart' = 7, 'MutatePartStart' = 8) COMMENT 'Type of the event that occurred with the data part. Can have one of the following values: NewPart — Inserting of a new data part, MergePartsStart — Merging of data parts has started, MergeParts — Merging of data parts has finished, DownloadPart — Downloading a data part, RemovePart — Removing or detaching a data part using [DETACH PARTITION](../../sql-reference/statements/alter/partition.md#alter_detach-partition).MutatePartStart — Mutating of a data part has started, MutatePart — Mutating of a data part has finished, MovePart — Moving the data part from the one disk to another one.', `merge_reason` Enum8('NotAMerge' = 1, 'RegularMerge' = 2, 'TTLDeleteMerge' = 3, 'TTLRecompressMerge' = 4) COMMENT 'The reason for the event with type MERGE_PARTS. Can have one of the following values: NotAMerge — The current event has the type other than MERGE_PARTS, RegularMerge — Some regular merge, TTLDeleteMerge — Cleaning up expired data. TTLRecompressMerge — Recompressing data part with the. ', `merge_algorithm` Enum8('Undecided' = 0, 'Vertical' = 1, 'Horizontal' = 2) COMMENT 'Merge algorithm for the event with type MERGE_PARTS. Can have one of the following values: Undecided, Horizontal, Vertical', `event_date` Date COMMENT 'Event date.', `event_time` DateTime COMMENT 'Event time.', `event_time_microseconds` DateTime64(6) COMMENT 'Event time with microseconds precision.', `duration_ms` UInt64 COMMENT 'Duration of this operation.', `database` String COMMENT 'Name of the database the data part is in.', `table` String COMMENT 'Name of the table the data part is in.', `table_uuid` UUID COMMENT 'UUID of the table the data part belongs to.', `part_name` String COMMENT 'Name of the data part.', `partition_id` String COMMENT 'ID of the partition that the data part was inserted to. The column takes the `all` value if the partitioning is by `tuple()`.', `partition` String COMMENT 'The partition name.', `part_type` String COMMENT 'The type of the part. Possible values: Wide and Compact.', `disk_name` String COMMENT 'The disk name data part lies on.', `path_on_disk` String COMMENT 'Absolute path to the folder with data part files.', `rows` UInt64 COMMENT 'The number of rows in the data part.', `size_in_bytes` UInt64 COMMENT 'Size of the data part on disk in bytes.', `merged_from` Array(String) COMMENT 'An array of the source parts names which the current part was made up from.', `bytes_uncompressed` UInt64 COMMENT 'Uncompressed size of the resulting part in bytes.', `read_rows` UInt64 COMMENT 'The number of rows was read during the merge.', `read_bytes` UInt64 COMMENT 'The number of bytes was read during the merge.', `peak_memory_usage` UInt64 COMMENT 'The maximum amount of used during merge RAM', `error` UInt16 COMMENT 'The error code of the occurred exception.', `exception` String COMMENT 'Text message of the occurred error.', `ProfileEvents` Map(LowCardinality(String), UInt64) COMMENT 'All the profile events captured during this operation.', `ProfileEvents.Names` Array(String) ALIAS mapKeys(ProfileEvents), `ProfileEvents.Values` Array(UInt64) ALIAS mapValues(ProfileEvents), `name` String ALIAS part_name ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'This table contains information about events that occurred with data parts in the MergeTree family tables, such as adding or merging data.\n\nIt is safe to truncate or drop this table at any time.' Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/456/testgarbage'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/456/testgarbage'] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/456/testgarbage] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/456/testgarbage] Executing query SHOW CREATE TABLE system.trace_log FORMAT TSVRaw on node_default Executing query SHOW CREATE TABLE system.trace_log FORMAT TSVRaw on node_default run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/456/30000000-1000-4000-8000-000000000003'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/456/30000000-1000-4000-8000-000000000003'] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/456/30000000-1000-4000-8000-000000000003] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/456/30000000-1000-4000-8000-000000000003] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['touch', '/var/lib/clickhouse//store/456/45600000-1000-4000-8000-000000000003'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['touch', '/var/lib/clickhouse//store/456/45600000-1000-4000-8000-000000000003'] Command:[docker exec rootteststorecleanup-gw1-node1-1 touch /var/lib/clickhouse//store/456/45600000-1000-4000-8000-000000000003] Command:[docker exec rootteststorecleanup-gw1-node1-1 touch /var/lib/clickhouse//store/456/45600000-1000-4000-8000-000000000003] With storage policy, SHOW CREATE TABLE system.trace_log is: CREATE TABLE system.trace_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `event_date` Date COMMENT 'Date of sampling moment.', `event_time` DateTime COMMENT 'Timestamp of the sampling moment.', `event_time_microseconds` DateTime64(6) COMMENT 'Timestamp of the sampling moment with microseconds precision.', `timestamp_ns` UInt64 COMMENT 'Timestamp of the sampling moment in nanoseconds.', `revision` UInt32 COMMENT 'ClickHouse server build revision.', `trace_type` Enum8('Real' = 0, 'CPU' = 1, 'Memory' = 2, 'MemorySample' = 3, 'MemoryPeak' = 4, 'ProfileEvent' = 5) COMMENT 'Trace type: `Real` represents collecting stack traces by wall-clock time. `CPU` represents collecting stack traces by CPU time. `Memory` represents collecting allocations and deallocations when memory allocation exceeds the subsequent watermark. `MemorySample` represents collecting random allocations and deallocations. `MemoryPeak` represents collecting updates of peak memory usage. `ProfileEvent` represents collecting of increments of profile events.', `thread_id` UInt64 COMMENT 'Thread identifier.', `query_id` String COMMENT 'Query identifier that can be used to get details about a query that was running from the query_log system table.', `trace` Array(UInt64) COMMENT 'Stack trace at the moment of sampling. Each element is a virtual memory address inside ClickHouse server process.', `size` Int64 COMMENT 'For trace types Memory, MemorySample or MemoryPeak is the amount of memory allocated, for other trace types is 0.', `ptr` UInt64 COMMENT 'The address of the allocated chunk.', `event` LowCardinality(String) COMMENT 'For trace type ProfileEvent is the name of updated profile event, for other trace types is an empty string.', `increment` Int64 COMMENT 'For trace type ProfileEvent is the amount of increment of profile event, for other trace types is 0.', `symbols` Array(LowCardinality(String)) COMMENT 'If the symbolization is enabled, contains demangled symbol names, corresponding to the `trace`.', `lines` Array(LowCardinality(String)) COMMENT 'If the symbolization is enabled, contains strings with file names with line numbers, corresponding to the `trace`.', `build_id` String ALIAS '4783FC52DD13D5DFB0294BDDD711047195FEB5A6' ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains stack traces collected by the sampling query profiler.\n\nIt is safe to truncate or drop this table at any time.' With storage policy, SHOW CREATE TABLE system.trace_log is: CREATE TABLE system.trace_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `event_date` Date COMMENT 'Date of sampling moment.', `event_time` DateTime COMMENT 'Timestamp of the sampling moment.', `event_time_microseconds` DateTime64(6) COMMENT 'Timestamp of the sampling moment with microseconds precision.', `timestamp_ns` UInt64 COMMENT 'Timestamp of the sampling moment in nanoseconds.', `revision` UInt32 COMMENT 'ClickHouse server build revision.', `trace_type` Enum8('Real' = 0, 'CPU' = 1, 'Memory' = 2, 'MemorySample' = 3, 'MemoryPeak' = 4, 'ProfileEvent' = 5) COMMENT 'Trace type: `Real` represents collecting stack traces by wall-clock time. `CPU` represents collecting stack traces by CPU time. `Memory` represents collecting allocations and deallocations when memory allocation exceeds the subsequent watermark. `MemorySample` represents collecting random allocations and deallocations. `MemoryPeak` represents collecting updates of peak memory usage. `ProfileEvent` represents collecting of increments of profile events.', `thread_id` UInt64 COMMENT 'Thread identifier.', `query_id` String COMMENT 'Query identifier that can be used to get details about a query that was running from the query_log system table.', `trace` Array(UInt64) COMMENT 'Stack trace at the moment of sampling. Each element is a virtual memory address inside ClickHouse server process.', `size` Int64 COMMENT 'For trace types Memory, MemorySample or MemoryPeak is the amount of memory allocated, for other trace types is 0.', `ptr` UInt64 COMMENT 'The address of the allocated chunk.', `event` LowCardinality(String) COMMENT 'For trace type ProfileEvent is the name of updated profile event, for other trace types is an empty string.', `increment` Int64 COMMENT 'For trace type ProfileEvent is the amount of increment of profile event, for other trace types is 0.', `symbols` Array(LowCardinality(String)) COMMENT 'If the symbolization is enabled, contains demangled symbol names, corresponding to the `trace`.', `lines` Array(LowCardinality(String)) COMMENT 'If the symbolization is enabled, contains strings with file names with line numbers, corresponding to the `trace`.', `build_id` String ALIAS '4783FC52DD13D5DFB0294BDDD711047195FEB5A6' ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains stack traces collected by the sampling query profiler.\n\nIt is safe to truncate or drop this table at any time.' Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/456/45600000-1000-4000-8000-000000000004'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['mkdir', '/var/lib/clickhouse//store/456/45600000-1000-4000-8000-000000000004'] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/456/45600000-1000-4000-8000-000000000004] Command:[docker exec rootteststorecleanup-gw1-node1-1 mkdir /var/lib/clickhouse//store/456/45600000-1000-4000-8000-000000000004] Executing query SHOW CREATE TABLE system.metric_log FORMAT TSVRaw on node_default Executing query SHOW CREATE TABLE system.metric_log FORMAT TSVRaw on node_default run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/rootteststorecleanup-gw1-node1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/rootteststorecleanup-gw1-node1-1/exec HTTP/1.1" 201 74 With storage policy, SHOW CREATE TABLE system.metric_log is: CREATE TABLE system.metric_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `event_date` Date COMMENT 'Event date.', `event_time` DateTime COMMENT 'Event time.', `event_time_microseconds` DateTime64(6) COMMENT 'Event time with microseconds resolution.', `ProfileEvent_Query` UInt64 COMMENT 'Number of queries to be interpreted and potentially executed. Does not include queries that failed to parse or were rejected due to AST size limits, quota limits or limits on the number of simultaneously running queries. May include internal queries initiated by ClickHouse itself. Does not count subqueries.', `ProfileEvent_SelectQuery` UInt64 COMMENT 'Same as Query, but only for SELECT queries.', `ProfileEvent_InsertQuery` UInt64 COMMENT 'Same as Query, but only for INSERT queries.', `ProfileEvent_InitialQuery` UInt64 COMMENT 'Same as Query, but only counts initial queries (see is_initial_query).', `ProfileEvent_QueriesWithSubqueries` UInt64 COMMENT 'Count queries with all subqueries', `ProfileEvent_SelectQueriesWithSubqueries` UInt64 COMMENT 'Count SELECT queries with all subqueries', `ProfileEvent_InsertQueriesWithSubqueries` UInt64 COMMENT 'Count INSERT queries with all subqueries', `ProfileEvent_SelectQueriesWithPrimaryKeyUsage` UInt64 COMMENT 'Count SELECT queries which use the primary key to evaluate the WHERE condition', `ProfileEvent_AsyncInsertQuery` UInt64 COMMENT 'Same as InsertQuery, but only for asynchronous INSERT queries.', `ProfileEvent_AsyncInsertBytes` UInt64 COMMENT 'Data size in bytes of asynchronous INSERT queries.', `ProfileEvent_AsyncInsertRows` UInt64 COMMENT 'Number of rows inserted by asynchronous INSERT queries.', `ProfileEvent_AsyncInsertCacheHits` UInt64 COMMENT 'Number of times a duplicate hash id has been found in asynchronous INSERT hash id cache.', `ProfileEvent_FailedQuery` UInt64 COMMENT 'Number of failed queries.', `ProfileEvent_FailedSelectQuery` UInt64 COMMENT 'Same as FailedQuery, but only for SELECT queries.', `ProfileEvent_FailedInsertQuery` UInt64 COMMENT 'Same as FailedQuery, but only for INSERT queries.', `ProfileEvent_FailedAsyncInsertQuery` UInt64 COMMENT 'Number of failed ASYNC INSERT queries.', `ProfileEvent_QueryTimeMicroseconds` UInt64 COMMENT 'Total time of all queries.', `ProfileEvent_SelectQueryTimeMicroseconds` UInt64 COMMENT 'Total time of SELECT queries.', `ProfileEvent_InsertQueryTimeMicroseconds` UInt64 COMMENT 'Total time of INSERT queries.', `ProfileEvent_OtherQueryTimeMicroseconds` UInt64 COMMENT 'Total time of queries that are not SELECT or INSERT.', `ProfileEvent_FileOpen` UInt64 COMMENT 'Number of files opened.', `ProfileEvent_Seek` UInt64 COMMENT 'Number of times the \'lseek\' function was called.', `ProfileEvent_ReadBufferFromFileDescriptorRead` UInt64 COMMENT 'Number of reads (read/pread) from a file descriptor. Does not include sockets.', `ProfileEvent_ReadBufferFromFileDescriptorReadFailed` UInt64 COMMENT 'Number of times the read (read/pread) from a file descriptor have failed.', `ProfileEvent_ReadBufferFromFileDescriptorReadBytes` UInt64 COMMENT 'Number of bytes read from file descriptors. If the file is compressed, this will show the compressed data size.', `ProfileEvent_WriteBufferFromFileDescriptorWrite` UInt64 COMMENT 'Number of writes (write/pwrite) to a file descriptor. Does not include sockets.', `ProfileEvent_WriteBufferFromFileDescriptorWriteFailed` UInt64 COMMENT 'Number of times the write (write/pwrite) to a file descriptor have failed.', `ProfileEvent_WriteBufferFromFileDescriptorWriteBytes` UInt64 COMMENT 'Number of bytes written to file descriptors. If the file is compressed, this will show compressed data size.', `ProfileEvent_FileSync` UInt64 COMMENT 'Number of times the F_FULLFSYNC/fsync/fdatasync function was called for files.', `ProfileEvent_DirectorySync` UInt64 COMMENT 'Number of times the F_FULLFSYNC/fsync/fdatasync function was called for directories.', `ProfileEvent_FileSyncElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for F_FULLFSYNC/fsync/fdatasync syscall for files.', `ProfileEvent_DirectorySyncElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for F_FULLFSYNC/fsync/fdatasync syscall for directories.', `ProfileEvent_ReadCompressedBytes` UInt64 COMMENT 'Number of bytes (the number of bytes before decompression) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferBlocks` UInt64 COMMENT 'Number of compressed blocks (the blocks of data that are compressed independent of each other) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferBytes` UInt64 COMMENT 'Number of uncompressed bytes (the number of bytes after decompression) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferChecksumDoesntMatch` UInt64 COMMENT 'Number of times the compressed block checksum did not match.', `ProfileEvent_CompressedReadBufferChecksumDoesntMatchSingleBitMismatch` UInt64 COMMENT 'Number of times a compressed block checksum mismatch was caused by a single-bit difference.', `ProfileEvent_CompressedReadBufferChecksumDoesntMatchMicroseconds` UInt64 COMMENT 'Total time spent detecting bit-flips due to compressed block checksum mismatches.', `ProfileEvent_UncompressedCacheHits` UInt64 COMMENT 'Number of times a block of data has been found in the uncompressed cache (and decompression was avoided).', `ProfileEvent_UncompressedCacheMisses` UInt64 COMMENT 'Number of times a block of data has not been found in the uncompressed cache (and required decompression).', `ProfileEvent_UncompressedCacheWeightLost` UInt64 COMMENT 'Number of bytes evicted from the uncompressed cache.', `ProfileEvent_MMappedFileCacheHits` UInt64 COMMENT 'Number of times a file has been found in the MMap cache (for the \'mmap\' read_method), so we didn\'t have to mmap it again.', `ProfileEvent_MMappedFileCacheMisses` UInt64 COMMENT 'Number of times a file has not been found in the MMap cache (for the \'mmap\' read_method), so we had to mmap it again.', `ProfileEvent_OpenedFileCacheHits` UInt64 COMMENT 'Number of times a file has been found in the opened file cache, so we didn\'t have to open it again.', `ProfileEvent_OpenedFileCacheMisses` UInt64 COMMENT 'Number of times a file has been found in the opened file cache, so we had to open it again.', `ProfileEvent_OpenedFileCacheMicroseconds` UInt64 COMMENT 'Amount of time spent executing OpenedFileCache methods.', `ProfileEvent_AIOWrite` UInt64 COMMENT 'Number of writes with Linux or FreeBSD AIO interface', `ProfileEvent_AIOWriteBytes` UInt64 COMMENT 'Number of bytes written with Linux or FreeBSD AIO interface', `ProfileEvent_AIORead` UInt64 COMMENT 'Number of reads with Linux or FreeBSD AIO interface', `ProfileEvent_AIOReadBytes` UInt64 COMMENT 'Number of bytes read with Linux or FreeBSD AIO interface', `ProfileEvent_IOBufferAllocs` UInt64 COMMENT 'Number of allocations of IO buffers (for ReadBuffer/WriteBuffer).', `ProfileEvent_IOBufferAllocBytes` UInt64 COMMENT 'Number of bytes allocated for IO buffers (for ReadBuffer/WriteBuffer).', `ProfileEvent_ArenaAllocChunks` UInt64 COMMENT 'Number of chunks allocated for memory Arena (used for GROUP BY and similar operations)', `ProfileEvent_ArenaAllocBytes` UInt64 COMMENT 'Number of bytes allocated for memory Arena (used for GROUP BY and similar operations)', `ProfileEvent_FunctionExecute` UInt64 COMMENT 'Number of SQL ordinary function calls (SQL functions are called on per-block basis, so this number represents the number of blocks).', `ProfileEvent_TableFunctionExecute` UInt64 COMMENT 'Number of table function calls.', `ProfileEvent_DefaultImplementationForNullsRows` UInt64 COMMENT 'Number of rows processed by default implementation for nulls in function execution', `ProfileEvent_DefaultImplementationForNullsRowsWithNulls` UInt64 COMMENT 'Number of rows which contain null values processed by default implementation for nulls in function execution', `ProfileEvent_MarkCacheHits` UInt64 COMMENT 'Number of times an entry has been found in the mark cache, so we didn\'t have to load a mark file.', `ProfileEvent_MarkCacheMisses` UInt64 COMMENT 'Number of times an entry has not been found in the mark cache, so we had to load a mark file in memory, which is a costly operation, adding to query latency.', `ProfileEvent_PrimaryIndexCacheHits` UInt64 COMMENT 'Number of times an entry has been found in the primary index cache, so we didn\'t have to load a index file.', `ProfileEvent_PrimaryIndexCacheMisses` UInt64 COMMENT 'Number of times an entry has not been found in the primary index cache, so we had to load a index file in memory, which is a costly operation, adding to query latency.', `ProfileEvent_SkippingIndexCacheHits` UInt64 COMMENT 'Number of times an index granule has been found in the skipping index cache.', `ProfileEvent_SkippingIndexCacheMisses` UInt64 COMMENT 'Number of times an index granule has not been found in the skipping index cache and had to be read from disk.', `ProfileEvent_SkippingIndexCacheWeightLost` UInt64 COMMENT 'Approximate number of bytes evicted from the secondary index cache.', `ProfileEvent_QueryCacheHits` UInt64 COMMENT 'Number of times a query result has been found in the query cache (and query computation was avoided). Only updated for SELECT queries with SETTING use_query_cache = 1.', `ProfileEvent_QueryCacheMisses` UInt64 COMMENT 'Number of times a query result has not been found in the query cache (and required query computation). Only updated for SELECT queries with SETTING use_query_cache = 1.', `ProfileEvent_PageCacheChunkMisses` UInt64 COMMENT 'Number of times a chunk has not been found in the userspace page cache.', `ProfileEvent_PageCacheChunkShared` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, already in use by another thread.', `ProfileEvent_PageCacheChunkDataHits` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, with all pages intact.', `ProfileEvent_PageCacheChunkDataPartialHits` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, but some of its pages were evicted by the OS.', `ProfileEvent_PageCacheChunkDataMisses` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, but all its pages were evicted by the OS.', `ProfileEvent_PageCacheBytesUnpinnedRoundedToPages` UInt64 COMMENT 'Total size of populated pages in chunks that became evictable in PageCache. Rounded up to whole pages.', `ProfileEvent_PageCacheBytesUnpinnedRoundedToHugePages` UInt64 COMMENT 'See PageCacheBytesUnpinnedRoundedToPages, but rounded to huge pages. Use the ratio between the two as a measure of memory waste from using huge pages.', `ProfileEvent_CreatedReadBufferOrdinary` UInt64 COMMENT 'Number of times ordinary read buffer was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferDirectIO` UInt64 COMMENT 'Number of times a read buffer with O_DIRECT was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferDirectIOFailed` UInt64 COMMENT 'Number of times a read buffer with O_DIRECT was attempted to be created for reading data (while choosing among other read methods), but the OS did not allow it (due to lack of filesystem support or other reasons) and we fallen back to the ordinary reading method.', `ProfileEvent_CreatedReadBufferMMap` UInt64 COMMENT 'Number of times a read buffer using \'mmap\' was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferMMapFailed` UInt64 COMMENT 'Number of times a read buffer with \'mmap\' was attempted to be created for reading data (while choosing among other read methods), but the OS did not allow it (due to lack of filesystem support or other reasons) and we fallen back to the ordinary reading method.', `ProfileEvent_DiskReadElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for read syscall. This include reads from page cache.', `ProfileEvent_DiskWriteElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for write syscall. This include writes to page cache.', `ProfileEvent_NetworkReceiveElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for data to receive or receiving data from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkSendElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for data to send to network or sending data to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkReceiveBytes` UInt64 COMMENT 'Total number of bytes received from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkSendBytes` UInt64 COMMENT 'Total number of bytes send to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_GlobalThreadPoolExpansions` UInt64 COMMENT 'Counts the total number of times new threads have been added to the global thread pool. This metric indicates the frequency of expansions in the global thread pool to accommodate increased processing demands.', `ProfileEvent_GlobalThreadPoolShrinks` UInt64 COMMENT 'Counts the total number of times the global thread pool has shrunk by removing threads. This occurs when the number of idle threads exceeds max_thread_pool_free_size, indicating adjustments in the global thread pool size in response to decreased thread utilization.', `ProfileEvent_GlobalThreadPoolThreadCreationMicroseconds` UInt64 COMMENT 'Total time spent waiting for new threads to start.', `ProfileEvent_GlobalThreadPoolLockWaitMicroseconds` UInt64 COMMENT 'Total time threads have spent waiting for locks in the global thread pool.', `ProfileEvent_GlobalThreadPoolJobs` UInt64 COMMENT 'Counts the number of jobs that have been pushed to the global thread pool.', `ProfileEvent_GlobalThreadPoolJobWaitTimeMicroseconds` UInt64 COMMENT 'Measures the elapsed time from when a job is scheduled in the thread pool to when it is picked up for execution by a worker thread. This metric helps identify delays in job processing, indicating the responsiveness of the thread pool to new tasks.', `ProfileEvent_LocalThreadPoolExpansions` UInt64 COMMENT 'Counts the total number of times threads have been borrowed from the global thread pool to expand local thread pools.', `ProfileEvent_LocalThreadPoolShrinks` UInt64 COMMENT 'Counts the total number of times threads have been returned to the global thread pool from local thread pools.', `ProfileEvent_LocalThreadPoolThreadCreationMicroseconds` UInt64 COMMENT 'Total time local thread pools have spent waiting to borrow a thread from the global pool.', `ProfileEvent_LocalThreadPoolLockWaitMicroseconds` UInt64 COMMENT 'Total time threads have spent waiting for locks in the local thread pools.', `ProfileEvent_LocalThreadPoolJobs` UInt64 COMMENT 'Counts the number of jobs that have been pushed to the local thread pools.', `ProfileEvent_LocalThreadPoolBusyMicroseconds` UInt64 COMMENT 'Total time threads have spent executing the actual work.', `ProfileEvent_LocalThreadPoolJobWaitTimeMicroseconds` UInt64 COMMENT 'Measures the elapsed time from when a job is scheduled in the thread pool to when it is picked up for execution by a worker thread. This metric helps identify delays in job processing, indicating the responsiveness of the thread pool to new tasks.', `ProfileEvent_DiskS3GetRequestThrottlerCount` UInt64 COMMENT 'Number of DiskS3 GET and SELECT requests passed through throttler.', `ProfileEvent_DiskS3GetRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform DiskS3 GET and SELECT request throttling.', `ProfileEvent_DiskS3PutRequestThrottlerCount` UInt64 COMMENT 'Number of DiskS3 PUT, COPY, POST and LIST requests passed through throttler.', `ProfileEvent_DiskS3PutRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform DiskS3 PUT, COPY, POST and LIST request throttling.', `ProfileEvent_S3GetRequestThrottlerCount` UInt64 COMMENT 'Number of S3 GET and SELECT requests passed through throttler.', `ProfileEvent_S3GetRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform S3 GET and SELECT request throttling.', `ProfileEvent_S3PutRequestThrottlerCount` UInt64 COMMENT 'Number of S3 PUT, COPY, POST and LIST requests passed through throttler.', `ProfileEvent_S3PutRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform S3 PUT, COPY, POST and LIST request throttling.', `ProfileEvent_RemoteReadThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_remote_read_network_bandwidth_for_server\'/\'max_remote_read_network_bandwidth\' throttler.', `ProfileEvent_RemoteReadThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_remote_read_network_bandwidth_for_server\'/\'max_remote_read_network_bandwidth\' throttling.', `ProfileEvent_RemoteWriteThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_remote_write_network_bandwidth_for_server\'/\'max_remote_write_network_bandwidth\' throttler.', `ProfileEvent_RemoteWriteThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_remote_write_network_bandwidth_for_server\'/\'max_remote_write_network_bandwidth\' throttling.', `ProfileEvent_LocalReadThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_local_read_bandwidth_for_server\'/\'max_local_read_bandwidth\' throttler.', `ProfileEvent_LocalReadThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_local_read_bandwidth_for_server\'/\'max_local_read_bandwidth\' throttling.', `ProfileEvent_LocalWriteThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_local_write_bandwidth_for_server\'/\'max_local_write_bandwidth\' throttler.', `ProfileEvent_LocalWriteThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_local_write_bandwidth_for_server\'/\'max_local_write_bandwidth\' throttling.', `ProfileEvent_ThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform all throttling settings.', `ProfileEvent_ReadTasksWithAppliedMutationsOnFly` UInt64 COMMENT 'Total number of parts for which there was any mutation applied on fly', `ProfileEvent_MutationsAppliedOnFlyInAllReadTasks` UInt64 COMMENT 'The sum of number of applied mutations on-fly for part among all read parts', `ProfileEvent_SchedulerIOReadRequests` UInt64 COMMENT 'Resource requests passed through scheduler for IO reads.', `ProfileEvent_SchedulerIOReadBytes` UInt64 COMMENT 'Bytes passed through scheduler for IO reads.', `ProfileEvent_SchedulerIOReadWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting on resource requests for IO reads.', `ProfileEvent_SchedulerIOWriteRequests` UInt64 COMMENT 'Resource requests passed through scheduler for IO writes.', `ProfileEvent_SchedulerIOWriteBytes` UInt64 COMMENT 'Bytes passed through scheduler for IO writes.', `ProfileEvent_SchedulerIOWriteWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting on resource requests for IO writes.', `ProfileEvent_QueryMaskingRulesMatch` UInt64 COMMENT 'Number of times query masking rules was successfully matched.', `ProfileEvent_ReplicatedPartFetches` UInt64 COMMENT 'Number of times a data part was downloaded from replica of a ReplicatedMergeTree table.', `ProfileEvent_ReplicatedPartFailedFetches` UInt64 COMMENT 'Number of times a data part was failed to download from replica of a ReplicatedMergeTree table.', `ProfileEvent_ObsoleteReplicatedParts` UInt64 COMMENT 'Number of times a data part was covered by another data part that has been fetched from a replica (so, we have marked a covered data part as obsolete and no longer needed).', `ProfileEvent_ReplicatedPartMerges` UInt64 COMMENT 'Number of times data parts of ReplicatedMergeTree tables were successfully merged.', `ProfileEvent_ReplicatedPartFetchesOfMerged` UInt64 COMMENT 'Number of times we prefer to download already merged part from replica of ReplicatedMergeTree table instead of performing a merge ourself (usually we prefer doing a merge ourself to save network traffic). This happens when we have not all source parts to perform a merge or when the data part is old enough.', `ProfileEvent_ReplicatedPartMutations` UInt64 COMMENT 'Number of times data parts of ReplicatedMergeTree tables were successfully mutated.', `ProfileEvent_ReplicatedPartChecks` UInt64 COMMENT 'Number of times we had to perform advanced search for a data part on replicas or to clarify the need of an existing data part.', `ProfileEvent_ReplicatedPartChecksFailed` UInt64 COMMENT 'Number of times the advanced search for a data part on replicas did not give result or when unexpected part has been found and moved away.', `ProfileEvent_ReplicatedDataLoss` UInt64 COMMENT 'Number of times a data part that we wanted doesn\'t exist on any replica (even on replicas that are offline right now). That data parts are definitely lost. This is normal due to asynchronous replication (if quorum inserts were not enabled), when the replica on which the data part was written was failed and when it became online after fail it doesn\'t contain that data part.', `ProfileEvent_ReplicatedCoveredPartsInZooKeeperOnStart` UInt64 COMMENT 'For debugging purposes. Number of parts in ZooKeeper that have a covering part, but doesn\'t exist on disk. Checked on server start.', `ProfileEvent_InsertedRows` UInt64 COMMENT 'Number of rows INSERTed to all tables.', `ProfileEvent_InsertedBytes` UInt64 COMMENT 'Number of bytes (uncompressed; for columns as they stored in memory) INSERTed to all tables.', `ProfileEvent_DelayedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a MergeTree table was throttled due to high number of active data parts for partition.', `ProfileEvent_RejectedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a MergeTree table was rejected with \'Too many parts\' exception due to high number of active data parts for partition.', `ProfileEvent_DelayedInsertsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the INSERT of a block to a MergeTree table was throttled due to high number of active data parts for partition.', `ProfileEvent_DelayedMutations` UInt64 COMMENT 'Number of times the mutation of a MergeTree table was throttled due to high number of unfinished mutations for table.', `ProfileEvent_RejectedMutations` UInt64 COMMENT 'Number of times the mutation of a MergeTree table was rejected with \'Too many mutations\' exception due to high number of unfinished mutations for table.', `ProfileEvent_DelayedMutationsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the mutation of a MergeTree table was throttled due to high number of unfinished mutations for table.', `ProfileEvent_DistributedDelayedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a Distributed table was throttled due to high number of pending bytes.', `ProfileEvent_DistributedRejectedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a Distributed table was rejected with \'Too many bytes\' exception due to high number of pending bytes.', `ProfileEvent_DistributedDelayedInsertsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the INSERT of a block to a Distributed table was throttled due to high number of pending bytes.', `ProfileEvent_DuplicatedInsertedBlocks` UInt64 COMMENT 'Number of times the INSERTed block to a ReplicatedMergeTree table was deduplicated.', `ProfileEvent_ZooKeeperInit` UInt64 COMMENT 'Number of times connection with ZooKeeper has been established.', `ProfileEvent_ZooKeeperTransactions` UInt64 COMMENT 'Number of ZooKeeper operations, which include both read and write operations as well as multi-transactions.', `ProfileEvent_ZooKeeperList` UInt64 COMMENT 'Number of \'list\' (getChildren) requests to ZooKeeper.', `ProfileEvent_ZooKeeperCreate` UInt64 COMMENT 'Number of \'create\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperRemove` UInt64 COMMENT 'Number of \'remove\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperExists` UInt64 COMMENT 'Number of \'exists\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperGet` UInt64 COMMENT 'Number of \'get\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperSet` UInt64 COMMENT 'Number of \'set\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperMulti` UInt64 COMMENT 'Number of \'multi\' requests to ZooKeeper (compound transactions).', `ProfileEvent_ZooKeeperCheck` UInt64 COMMENT 'Number of \'check\' requests to ZooKeeper. Usually they don\'t make sense in isolation, only as part of a complex transaction.', `ProfileEvent_ZooKeeperSync` UInt64 COMMENT 'Number of \'sync\' requests to ZooKeeper. These requests are rarely needed or usable.', `ProfileEvent_ZooKeeperReconfig` UInt64 COMMENT 'Number of \'reconfig\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperClose` UInt64 COMMENT 'Number of times connection with ZooKeeper has been closed voluntary.', `ProfileEvent_ZooKeeperWatchResponse` UInt64 COMMENT 'Number of times watch notification has been received from ZooKeeper.', `ProfileEvent_ZooKeeperUserExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper related to the data (no node, bad version or similar).', `ProfileEvent_ZooKeeperHardwareExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper related to network (connection loss or similar).', `ProfileEvent_ZooKeeperOtherExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper other than ZooKeeperUserExceptions and ZooKeeperHardwareExceptions.', `ProfileEvent_ZooKeeperWaitMicroseconds` UInt64 COMMENT 'Number of microseconds spent waiting for responses from ZooKeeper after creating a request, summed across all the requesting threads.', `ProfileEvent_ZooKeeperBytesSent` UInt64 COMMENT 'Number of bytes send over network while communicating with ZooKeeper.', `ProfileEvent_ZooKeeperBytesReceived` UInt64 COMMENT 'Number of bytes received over network while communicating with ZooKeeper.', `ProfileEvent_DistributedConnectionTries` UInt64 COMMENT 'Total count of distributed connection attempts.', `ProfileEvent_DistributedConnectionUsable` UInt64 COMMENT 'Total count of successful distributed connections to a usable server (with required table, but maybe stale).', `ProfileEvent_DistributedConnectionFailTry` UInt64 COMMENT 'Total count when distributed connection fails with retry.', `ProfileEvent_DistributedConnectionMissingTable` UInt64 COMMENT 'Number of times we rejected a replica from a distributed query, because it did not contain a table needed for the query.', `ProfileEvent_DistributedConnectionStaleReplica` UInt64 COMMENT 'Number of times we rejected a replica from a distributed query, because some table needed for a query had replication lag higher than the configured threshold.', `ProfileEvent_DistributedConnectionSkipReadOnlyReplica` UInt64 COMMENT 'Number of replicas skipped during INSERT into Distributed table due to replicas being read-only', `ProfileEvent_DistributedConnectionFailAtAll` UInt64 COMMENT 'Total count when distributed connection fails after all retries finished.', `ProfileEvent_HedgedRequestsChangeReplica` UInt64 COMMENT 'Total count when timeout for changing replica expired in hedged requests.', `ProfileEvent_SuspendSendingQueryToShard` UInt64 COMMENT 'Total count when sending query to shard was suspended when async_query_sending_for_remote is enabled.', `ProfileEvent_CompileFunction` UInt64 COMMENT 'Number of times a compilation of generated LLVM code (to create fused function for complex expressions) was initiated.', `ProfileEvent_CompiledFunctionExecute` UInt64 COMMENT 'Number of times a compiled function was executed.', `ProfileEvent_CompileExpressionsMicroseconds` UInt64 COMMENT 'Total time spent for compilation of expressions to LLVM code.', `ProfileEvent_CompileExpressionsBytes` UInt64 COMMENT 'Number of bytes used for expressions compilation.', `ProfileEvent_ExecuteShellCommand` UInt64 COMMENT 'Number of shell command executions.', `ProfileEvent_ExternalProcessingCompressedBytesTotal` UInt64 COMMENT 'Number of compressed bytes written by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalProcessingUncompressedBytesTotal` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalProcessingFilesTotal` UInt64 COMMENT 'Number of files used by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalSortWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for sorting in external memory.', `ProfileEvent_ExternalSortMerge` UInt64 COMMENT 'Number of times temporary files were merged for sorting in external memory.', `ProfileEvent_ExternalSortCompressedBytes` UInt64 COMMENT 'Number of compressed bytes written for sorting in external memory.', `ProfileEvent_ExternalSortUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written for sorting in external memory.', `ProfileEvent_ExternalAggregationWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for aggregation in external memory.', `ProfileEvent_ExternalAggregationMerge` UInt64 COMMENT 'Number of times temporary files were merged for aggregation in external memory.', `ProfileEvent_ExternalAggregationCompressedBytes` UInt64 COMMENT 'Number of bytes written to disk for aggregation in external memory.', `ProfileEvent_ExternalAggregationUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written to disk for aggregation in external memory.', `ProfileEvent_ExternalJoinWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for JOIN in external memory.', `ProfileEvent_ExternalJoinMerge` UInt64 COMMENT 'Number of times temporary files were merged for JOIN in external memory.', `ProfileEvent_ExternalJoinCompressedBytes` UInt64 COMMENT 'Number of compressed bytes written for JOIN in external memory.', `ProfileEvent_ExternalJoinUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written for JOIN in external memory.', `ProfileEvent_IcebergPartitionPrunnedFiles` UInt64 COMMENT 'Number of skipped files during Iceberg partition pruning', `ProfileEvent_JoinBuildTableRowCount` UInt64 COMMENT 'Total number of rows in the build table for a JOIN operation.', `ProfileEvent_JoinProbeTableRowCount` UInt64 COMMENT 'Total number of rows in the probe table for a JOIN operation.', `ProfileEvent_JoinResultRowCount` UInt64 COMMENT 'Total number of rows in the result of a JOIN operation.', `ProfileEvent_SlowRead` UInt64 COMMENT 'Number of reads from a file that were slow. This indicate system overload. Thresholds are controlled by read_backoff_* settings.', `ProfileEvent_ReadBackoff` UInt64 COMMENT 'Number of times the number of query processing threads was lowered due to slow reads.', `ProfileEvent_ReplicaPartialShutdown` UInt64 COMMENT 'How many times Replicated table has to deinitialize its state due to session expiration in ZooKeeper. The state is reinitialized every time when ZooKeeper is available again.', `ProfileEvent_SelectedParts` UInt64 COMMENT 'Number of data parts selected to read from a MergeTree table.', `ProfileEvent_SelectedPartsTotal` UInt64 COMMENT 'Number of total data parts before selecting which ones to read from a MergeTree table.', `ProfileEvent_SelectedRanges` UInt64 COMMENT 'Number of (non-adjacent) ranges in all data parts selected to read from a MergeTree table.', `ProfileEvent_SelectedMarks` UInt64 COMMENT 'Number of marks (index granules) selected to read from a MergeTree table.', `ProfileEvent_SelectedMarksTotal` UInt64 COMMENT 'Number of total marks (index granules) before selecting which ones to read from a MergeTree table.', `ProfileEvent_SelectedRows` UInt64 COMMENT 'Number of rows SELECTed from all tables.', `ProfileEvent_SelectedBytes` UInt64 COMMENT 'Number of bytes (uncompressed; for columns as they stored in memory) SELECTed from all tables.', `ProfileEvent_RowsReadByMainReader` UInt64 COMMENT 'Number of rows read from MergeTree tables by the main reader (after PREWHERE step).', `ProfileEvent_RowsReadByPrewhereReaders` UInt64 COMMENT 'Number of rows read from MergeTree tables (in total) by prewhere readers.', `ProfileEvent_LoadedDataParts` UInt64 COMMENT 'Number of data parts loaded by MergeTree tables during initialization.', `ProfileEvent_LoadedDataPartsMicroseconds` UInt64 COMMENT 'Microseconds spent by MergeTree tables for loading data parts during initialization.', `ProfileEvent_WaitMarksLoadMicroseconds` UInt64 COMMENT 'Time spent loading marks', `ProfileEvent_BackgroundLoadingMarksTasks` UInt64 COMMENT 'Number of background tasks for loading marks', `ProfileEvent_LoadingMarksTasksCanceled` UInt64 COMMENT 'Number of times background tasks for loading marks were canceled', `ProfileEvent_LoadedMarksFiles` UInt64 COMMENT 'Number of mark files loaded.', `ProfileEvent_LoadedMarksCount` UInt64 COMMENT 'Number of marks loaded (total across columns).', `ProfileEvent_LoadedMarksMemoryBytes` UInt64 COMMENT 'Size of in-memory representations of loaded marks.', `ProfileEvent_LoadedPrimaryIndexFiles` UInt64 COMMENT 'Number of primary index files loaded.', `ProfileEvent_LoadedPrimaryIndexRows` UInt64 COMMENT 'Number of rows of primary key loaded.', `ProfileEvent_LoadedPrimaryIndexBytes` UInt64 COMMENT 'Number of rows of primary key loaded.', `ProfileEvent_Merge` UInt64 COMMENT 'Number of launched background merges.', `ProfileEvent_MergeSourceParts` UInt64 COMMENT 'Number of source parts scheduled for merges.', `ProfileEvent_MergedRows` UInt64 COMMENT 'Rows read for background merges. This is the number of rows before merge.', `ProfileEvent_MergedColumns` UInt64 COMMENT 'Number of columns merged during the horizontal stage of merges.', `ProfileEvent_GatheredColumns` UInt64 COMMENT 'Number of columns gathered during the vertical stage of merges.', `ProfileEvent_MergedUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) that was read for background merges. This is the number before merge.', `ProfileEvent_MergeTotalMilliseconds` UInt64 COMMENT 'Total time spent for background merges', `ProfileEvent_MergeExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of background merges', `ProfileEvent_MergeHorizontalStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for horizontal stage of background merges', `ProfileEvent_MergeHorizontalStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of horizontal stage of background merges', `ProfileEvent_MergeVerticalStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for vertical stage of background merges', `ProfileEvent_MergeVerticalStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of vertical stage of background merges', `ProfileEvent_MergeProjectionStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for projection stage of background merges', `ProfileEvent_MergeProjectionStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of projection stage of background merges', `ProfileEvent_MergePrewarmStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for prewarm stage of background merges', `ProfileEvent_MergePrewarmStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of prewarm stage of background merges', `ProfileEvent_MergingSortedMilliseconds` UInt64 COMMENT 'Total time spent while merging sorted columns', `ProfileEvent_AggregatingSortedMilliseconds` UInt64 COMMENT 'Total time spent while aggregating sorted columns', `ProfileEvent_CollapsingSortedMilliseconds` UInt64 COMMENT 'Total time spent while collapsing sorted columns', `ProfileEvent_ReplacingSortedMilliseconds` UInt64 COMMENT 'Total time spent while replacing sorted columns', `ProfileEvent_SummingSortedMilliseconds` UInt64 COMMENT 'Total time spent while summing sorted columns', `ProfileEvent_VersionedCollapsingSortedMilliseconds` UInt64 COMMENT 'Total time spent while version collapsing sorted columns', `ProfileEvent_GatheringColumnMilliseconds` UInt64 COMMENT 'Total time spent while gathering columns for vertical merge', `ProfileEvent_MutationTotalParts` UInt64 COMMENT 'Number of total parts for which mutations tried to be applied', `ProfileEvent_MutationUntouchedParts` UInt64 COMMENT 'Number of total parts for which mutations tried to be applied but which was completely skipped according to predicate', `ProfileEvent_MutatedRows` UInt64 COMMENT 'Rows read for mutations. This is the number of rows before mutation', `ProfileEvent_MutatedUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) that was read for mutations. This is the number before mutation.', `ProfileEvent_MutationTotalMilliseconds` UInt64 COMMENT 'Total time spent for mutations.', `ProfileEvent_MutationExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of mutations.', `ProfileEvent_MutationAllPartColumns` UInt64 COMMENT 'Number of times when task to mutate all columns in part was created', `ProfileEvent_MutationSomePartColumns` UInt64 COMMENT 'Number of times when task to mutate some columns in part was created', `ProfileEvent_MutateTaskProjectionsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating projections in mutations', `ProfileEvent_MergeTreeDataWriterRows` UInt64 COMMENT 'Number of rows INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterCompressedBytes` UInt64 COMMENT 'Bytes written to filesystem for data INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterBlocks` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables. Each block forms a data part of level zero.', `ProfileEvent_MergeTreeDataWriterBlocksAlreadySorted` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables that appeared to be already sorted.', `ProfileEvent_MergeTreeDataWriterSkipIndicesCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating skip indices', `ProfileEvent_MergeTreeDataWriterStatisticsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating statistics', `ProfileEvent_MergeTreeDataWriterSortingBlocksMicroseconds` UInt64 COMMENT 'Time spent sorting blocks', `ProfileEvent_MergeTreeDataWriterMergingBlocksMicroseconds` UInt64 COMMENT 'Time spent merging input blocks (for special MergeTree engines)', `ProfileEvent_MergeTreeDataWriterProjectionsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating projections', `ProfileEvent_MergeTreeDataProjectionWriterSortingBlocksMicroseconds` UInt64 COMMENT 'Time spent sorting blocks (for projection it might be a key different from table\'s sorting key)', `ProfileEvent_MergeTreeDataProjectionWriterMergingBlocksMicroseconds` UInt64 COMMENT 'Time spent merging blocks', `ProfileEvent_InsertedWideParts` UInt64 COMMENT 'Number of parts inserted in Wide format.', `ProfileEvent_InsertedCompactParts` UInt64 COMMENT 'Number of parts inserted in Compact format.', `ProfileEvent_MergedIntoWideParts` UInt64 COMMENT 'Number of parts merged into Wide format.', `ProfileEvent_MergedIntoCompactParts` UInt64 COMMENT 'Number of parts merged into Compact format.', `ProfileEvent_MergeTreeDataProjectionWriterRows` UInt64 COMMENT 'Number of rows INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterCompressedBytes` UInt64 COMMENT 'Bytes written to filesystem for data INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterBlocks` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables projection. Each block forms a data part of level zero.', `ProfileEvent_MergeTreeDataProjectionWriterBlocksAlreadySorted` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables projection that appeared to be already sorted.', `ProfileEvent_CannotRemoveEphemeralNode` UInt64 COMMENT 'Number of times an error happened while trying to remove ephemeral node. This is not an issue, because our implementation of ZooKeeper library guarantee that the session will expire and the node will be removed.', `ProfileEvent_RegexpWithMultipleNeedlesCreated` UInt64 COMMENT 'Regular expressions with multiple needles (VectorScan library) compiled.', `ProfileEvent_RegexpWithMultipleNeedlesGlobalCacheHit` UInt64 COMMENT 'Number of times we fetched compiled regular expression with multiple needles (VectorScan library) from the global cache.', `ProfileEvent_RegexpWithMultipleNeedlesGlobalCacheMiss` UInt64 COMMENT 'Number of times we failed to fetch compiled regular expression with multiple needles (VectorScan library) from the global cache.', `ProfileEvent_RegexpLocalCacheHit` UInt64 COMMENT 'Number of times we fetched compiled regular expression from a local cache.', `ProfileEvent_RegexpLocalCacheMiss` UInt64 COMMENT 'Number of times we failed to fetch compiled regular expression from a local cache.', `ProfileEvent_ContextLock` UInt64 COMMENT 'Number of times the lock of Context was acquired or tried to acquire. This is global lock.', `ProfileEvent_ContextLockWaitMicroseconds` UInt64 COMMENT 'Context lock wait time in microseconds', `ProfileEvent_StorageBufferFlush` UInt64 COMMENT 'Number of times a buffer in a \'Buffer\' table was flushed.', `ProfileEvent_StorageBufferErrorOnFlush` UInt64 COMMENT 'Number of times a buffer in the \'Buffer\' table has not been able to flush due to error writing in the destination table.', `ProfileEvent_StorageBufferPassedAllMinThresholds` UInt64 COMMENT 'Number of times a criteria on min thresholds has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedTimeMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max time threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedRowsMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max rows threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedBytesMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max bytes threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedTimeFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on time has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferPassedRowsFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on rows has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferPassedBytesFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on bytes has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferLayerLockReadersWaitMilliseconds` UInt64 COMMENT 'Time for waiting for Buffer layer during reading.', `ProfileEvent_StorageBufferLayerLockWritersWaitMilliseconds` UInt64 COMMENT 'Time for waiting free Buffer layer to write to (can be used to tune Buffer layers).', `ProfileEvent_SystemLogErrorOnFlush` UInt64 COMMENT 'Number of times any of the system logs have failed to flush to the corresponding system table. Attempts to flush are repeated.', `ProfileEvent_DictCacheKeysRequested` UInt64 COMMENT 'Number of keys requested from the data source for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheKeysRequestedMiss` UInt64 COMMENT 'Number of keys requested from the data source for dictionaries of \'cache\' types but not found in the data source.', `ProfileEvent_DictCacheKeysRequestedFound` UInt64 COMMENT 'Number of keys requested from the data source for dictionaries of \'cache\' types and found in the data source.', `ProfileEvent_DictCacheKeysExpired` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and found in the cache but they were obsolete.', `ProfileEvent_DictCacheKeysNotFound` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and not found.', `ProfileEvent_DictCacheKeysHit` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and found in the cache.', `ProfileEvent_DictCacheRequestTimeNs` UInt64 COMMENT 'Number of nanoseconds spend in querying the external data sources for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheRequests` UInt64 COMMENT 'Number of bulk requests to the external data sources for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheLockWriteNs` UInt64 COMMENT 'Number of nanoseconds spend in waiting for write lock to update the data for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheLockReadNs` UInt64 COMMENT 'Number of nanoseconds spend in waiting for read lock to lookup the data for the dictionaries of \'cache\' types.', `ProfileEvent_DistributedSyncInsertionTimeoutExceeded` UInt64 COMMENT 'A timeout has exceeded while waiting for shards during synchronous insertion into a Distributed table (with \'distributed_foreground_insert\' = 1)', `ProfileEvent_DistributedAsyncInsertionFailures` UInt64 COMMENT 'Number of failures for asynchronous insertion into a Distributed table (with \'distributed_foreground_insert\' = 0)', `ProfileEvent_DataAfterMergeDiffersFromReplica` UInt64 COMMENT '\nNumber of times data after merge is not byte-identical to the data on another replicas. There could be several reasons:\n1. Using newer version of compression library after server update.\n2. Using another compression method.\n3. Non-deterministic compression algorithm (highly unlikely).\n4. Non-deterministic merge algorithm due to logical error in code.\n5. Data corruption in memory due to bug in code.\n6. Data corruption in memory due to hardware issue.\n7. Manual modification of source data after server startup.\n8. Manual modification of checksums stored in ZooKeeper.\n9. Part format related settings like \'enable_mixed_granularity_parts\' are different on different replicas.\nThe server successfully detected this situation and will download merged part from the replica to force the byte-identical result.\n', `ProfileEvent_DataAfterMutationDiffersFromReplica` UInt64 COMMENT 'Number of times data after mutation is not byte-identical to the data on other replicas. In addition to the reasons described in \'DataAfterMergeDiffersFromReplica\', it is also possible due to non-deterministic mutation.', `ProfileEvent_PolygonsAddedToPool` UInt64 COMMENT 'A polygon has been added to the cache (pool) for the \'pointInPolygon\' function.', `ProfileEvent_PolygonsInPoolAllocatedBytes` UInt64 COMMENT 'The number of bytes for polygons added to the cache (pool) for the \'pointInPolygon\' function.', `ProfileEvent_USearchAddCount` UInt64 COMMENT 'Number of vectors added to usearch indexes.', `ProfileEvent_USearchAddVisitedMembers` UInt64 COMMENT 'Number of nodes visited when adding vectors to usearch indexes.', `ProfileEvent_USearchAddComputedDistances` UInt64 COMMENT 'Number of times distance was computed when adding vectors to usearch indexes.', `ProfileEvent_USearchSearchCount` UInt64 COMMENT 'Number of search operations performed in usearch indexes.', `ProfileEvent_USearchSearchVisitedMembers` UInt64 COMMENT 'Number of nodes visited when searching in usearch indexes.', `ProfileEvent_USearchSearchComputedDistances` UInt64 COMMENT 'Number of times distance was computed when searching usearch indexes.', `ProfileEvent_RWLockAcquiredReadLocks` UInt64 COMMENT 'Number of times a read lock was acquired (in a heavy RWLock).', `ProfileEvent_RWLockAcquiredWriteLocks` UInt64 COMMENT 'Number of times a write lock was acquired (in a heavy RWLock).', `ProfileEvent_RWLockReadersWaitMilliseconds` UInt64 COMMENT 'Total time spent waiting for a read lock to be acquired (in a heavy RWLock).', `ProfileEvent_RWLockWritersWaitMilliseconds` UInt64 COMMENT 'Total time spent waiting for a write lock to be acquired (in a heavy RWLock).', `ProfileEvent_DNSError` UInt64 COMMENT 'Total count of errors in DNS resolution', `ProfileEvent_PartsLockHoldMicroseconds` UInt64 COMMENT 'Total time spent holding data parts lock in MergeTree tables', `ProfileEvent_PartsLockWaitMicroseconds` UInt64 COMMENT 'Total time spent waiting for data parts lock in MergeTree tables', `ProfileEvent_RealTimeMicroseconds` UInt64 COMMENT 'Total (wall clock) time spent in processing (queries and other tasks) threads (note that this is a sum).', `ProfileEvent_UserTimeMicroseconds` UInt64 COMMENT 'Total time spent in processing (queries and other tasks) threads executing CPU instructions in user mode. This includes time CPU pipeline was stalled due to main memory access, cache misses, branch mispredictions, hyper-threading, etc.', `ProfileEvent_SystemTimeMicroseconds` UInt64 COMMENT 'Total time spent in processing (queries and other tasks) threads executing CPU instructions in OS kernel mode. This is time spent in syscalls, excluding waiting time during blocking syscalls.', `ProfileEvent_MemoryOvercommitWaitTimeMicroseconds` UInt64 COMMENT 'Total time spent in waiting for memory to be freed in OvercommitTracker.', `ProfileEvent_MemoryAllocatorPurge` UInt64 COMMENT 'Total number of times memory allocator purge was requested', `ProfileEvent_MemoryAllocatorPurgeTimeMicroseconds` UInt64 COMMENT 'Total number of times memory allocator purge was requested', `ProfileEvent_SoftPageFaults` UInt64 COMMENT 'The number of soft page faults in query execution threads. Soft page fault usually means a miss in the memory allocator cache, which requires a new memory mapping from the OS and subsequent allocation of a page of physical memory.', `ProfileEvent_HardPageFaults` UInt64 COMMENT 'The number of hard page faults in query execution threads. High values indicate either that you forgot to turn off swap on your server, or eviction of memory pages of the ClickHouse binary during very high memory pressure, or successful usage of the \'mmap\' read method for the tables data.', `ProfileEvent_OSIOWaitMicroseconds` UInt64 COMMENT 'Total time a thread spent waiting for a result of IO operation, from the OS point of view. This is real IO that doesn\'t include page cache.', `ProfileEvent_OSCPUWaitMicroseconds` UInt64 COMMENT 'Total time a thread was ready for execution but waiting to be scheduled by OS, from the OS point of view.', `ProfileEvent_OSCPUVirtualTimeMicroseconds` UInt64 COMMENT 'CPU time spent seen by OS. Does not include involuntary waits due to virtualization.', `ProfileEvent_OSReadBytes` UInt64 COMMENT 'Number of bytes read from disks or block devices. Doesn\'t include bytes read from page cache. May include excessive data due to block size, readahead, etc.', `ProfileEvent_OSWriteBytes` UInt64 COMMENT 'Number of bytes written to disks or block devices. Doesn\'t include bytes that are in page cache dirty pages. May not include data that was written by OS asynchronously.', `ProfileEvent_OSReadChars` UInt64 COMMENT 'Number of bytes read from filesystem, including page cache.', `ProfileEvent_OSWriteChars` UInt64 COMMENT 'Number of bytes written to filesystem, including page cache.', `ProfileEvent_ParallelReplicasHandleRequestMicroseconds` UInt64 COMMENT 'Time spent processing requests for marks from replicas', `ProfileEvent_ParallelReplicasHandleAnnouncementMicroseconds` UInt64 COMMENT 'Time spent processing replicas announcements', `ProfileEvent_ParallelReplicasAnnouncementMicroseconds` UInt64 COMMENT 'Time spent to send an announcement', `ProfileEvent_ParallelReplicasReadRequestMicroseconds` UInt64 COMMENT 'Time spent for read requests', `ProfileEvent_ParallelReplicasReadAssignedMarks` UInt64 COMMENT 'Sum across all replicas of how many of scheduled marks were assigned by consistent hash', `ProfileEvent_ParallelReplicasReadUnassignedMarks` UInt64 COMMENT 'Sum across all replicas of how many unassigned marks were scheduled', `ProfileEvent_ParallelReplicasReadAssignedForStealingMarks` UInt64 COMMENT 'Sum across all replicas of how many of scheduled marks were assigned for stealing by consistent hash', `ProfileEvent_ParallelReplicasReadMarks` UInt64 COMMENT 'How many marks were read by the given replica', `ProfileEvent_ParallelReplicasStealingByHashMicroseconds` UInt64 COMMENT 'Time spent collecting segments meant for stealing by hash', `ProfileEvent_ParallelReplicasProcessingPartsMicroseconds` UInt64 COMMENT 'Time spent processing data parts', `ProfileEvent_ParallelReplicasStealingLeftoversMicroseconds` UInt64 COMMENT 'Time spent collecting orphaned segments', `ProfileEvent_ParallelReplicasCollectingOwnedSegmentsMicroseconds` UInt64 COMMENT 'Time spent collecting segments meant by hash', `ProfileEvent_ParallelReplicasNumRequests` UInt64 COMMENT 'Number of requests to the initiator.', `ProfileEvent_ParallelReplicasDeniedRequests` UInt64 COMMENT 'Number of completely denied requests to the initiator', `ProfileEvent_CacheWarmerBytesDownloaded` UInt64 COMMENT 'Amount of data fetched into filesystem cache by dedicated background threads.', `ProfileEvent_CacheWarmerDataPartsDownloaded` UInt64 COMMENT 'Number of data parts that were fully fetched by CacheWarmer.', `ProfileEvent_IgnoredColdParts` UInt64 COMMENT 'See setting ignore_cold_parts_seconds. Number of times read queries ignored very new parts that weren\'t pulled into cache by CacheWarmer yet.', `ProfileEvent_PreferredWarmedUnmergedParts` UInt64 COMMENT 'See setting prefer_warmed_unmerged_parts_seconds. Number of times read queries used outdated pre-merge parts that are in cache instead of merged part that wasn\'t pulled into cache by CacheWarmer yet.', `ProfileEvent_PerfCPUCycles` UInt64 COMMENT 'Total cycles. Be wary of what happens during CPU frequency scaling.', `ProfileEvent_PerfInstructions` UInt64 COMMENT 'Retired instructions. Be careful, these can be affected by various issues, most notably hardware interrupt counts.', `ProfileEvent_PerfCacheReferences` UInt64 COMMENT 'Cache accesses. Usually, this indicates Last Level Cache accesses, but this may vary depending on your CPU. This may include prefetches and coherency messages; again this depends on the design of your CPU.', `ProfileEvent_PerfCacheMisses` UInt64 COMMENT 'Cache misses. Usually this indicates Last Level Cache misses; this is intended to be used in conjunction with the PERFCOUNTHWCACHEREFERENCES event to calculate cache miss rates.', `ProfileEvent_PerfBranchInstructions` UInt64 COMMENT 'Retired branch instructions. Prior to Linux 2.6.35, this used the wrong event on AMD processors.', `ProfileEvent_PerfBranchMisses` UInt64 COMMENT 'Mispredicted branch instructions.', `ProfileEvent_PerfBusCycles` UInt64 COMMENT 'Bus cycles, which can be different from total cycles.', `ProfileEvent_PerfStalledCyclesFrontend` UInt64 COMMENT 'Stalled cycles during issue.', `ProfileEvent_PerfStalledCyclesBackend` UInt64 COMMENT 'Stalled cycles during retirement.', `ProfileEvent_PerfRefCPUCycles` UInt64 COMMENT 'Total cycles; not affected by CPU frequency scaling.', `ProfileEvent_PerfCPUClock` UInt64 COMMENT 'The CPU clock, a high-resolution per-CPU timer', `ProfileEvent_PerfTaskClock` UInt64 COMMENT 'A clock count specific to the task that is running', `ProfileEvent_PerfContextSwitches` UInt64 COMMENT 'Number of context switches', `ProfileEvent_PerfCPUMigrations` UInt64 COMMENT 'Number of times the process has migrated to a new CPU', `ProfileEvent_PerfAlignmentFaults` UInt64 COMMENT 'Number of alignment faults. These happen when unaligned memory accesses happen; the kernel can handle these but it reduces performance. This happens only on some architectures (never on x86).', `ProfileEvent_PerfEmulationFaults` UInt64 COMMENT 'Number of emulation faults. The kernel sometimes traps on unimplemented instructions and emulates them for user space. This can negatively impact performance.', `ProfileEvent_PerfMinEnabledTime` UInt64 COMMENT 'For all events, minimum time that an event was enabled. Used to track event multiplexing influence', `ProfileEvent_PerfMinEnabledRunningTime` UInt64 COMMENT 'Running time for event with minimum enabled time. Used to track the amount of event multiplexing', `ProfileEvent_PerfDataTLBReferences` UInt64 COMMENT 'Data TLB references', `ProfileEvent_PerfDataTLBMisses` UInt64 COMMENT 'Data TLB misses', `ProfileEvent_PerfInstructionTLBReferences` UInt64 COMMENT 'Instruction TLB references', `ProfileEvent_PerfInstructionTLBMisses` UInt64 COMMENT 'Instruction TLB misses', `ProfileEvent_PerfLocalMemoryReferences` UInt64 COMMENT 'Local NUMA node memory reads', `ProfileEvent_PerfLocalMemoryMisses` UInt64 COMMENT 'Local NUMA node memory read misses', `ProfileEvent_CannotWriteToWriteBufferDiscard` UInt64 COMMENT 'Number of stack traces dropped by query profiler or signal handler because pipe is full or cannot write to pipe.', `ProfileEvent_QueryProfilerSignalOverruns` UInt64 COMMENT 'Number of times we drop processing of a query profiler signal due to overrun plus the number of signals that OS has not delivered due to overrun.', `ProfileEvent_QueryProfilerConcurrencyOverruns` UInt64 COMMENT 'Number of times we drop processing of a query profiler signal due to too many concurrent query profilers in other threads, which may indicate overload.', `ProfileEvent_QueryProfilerRuns` UInt64 COMMENT 'Number of times QueryProfiler had been run.', `ProfileEvent_QueryProfilerErrors` UInt64 COMMENT 'Invalid memory accesses during asynchronous stack unwinding.', `ProfileEvent_CreatedLogEntryForMerge` UInt64 COMMENT 'Successfully created log entry to merge parts in ReplicatedMergeTree.', `ProfileEvent_NotCreatedLogEntryForMerge` UInt64 COMMENT 'Log entry to merge parts in ReplicatedMergeTree is not created due to concurrent log update by another replica.', `ProfileEvent_CreatedLogEntryForMutation` UInt64 COMMENT 'Successfully created log entry to mutate parts in ReplicatedMergeTree.', `ProfileEvent_NotCreatedLogEntryForMutation` UInt64 COMMENT 'Log entry to mutate parts in ReplicatedMergeTree is not created due to concurrent log update by another replica.', `ProfileEvent_S3ReadMicroseconds` UInt64 COMMENT 'Time of GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsCount` UInt64 COMMENT 'Number of GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsRedirects` UInt64 COMMENT 'Number of redirects in GET and HEAD requests to S3 storage.', `ProfileEvent_S3WriteMicroseconds` UInt64 COMMENT 'Time of POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsCount` UInt64 COMMENT 'Number of POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsRedirects` UInt64 COMMENT 'Number of redirects in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_DiskS3ReadMicroseconds` UInt64 COMMENT 'Time of GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsCount` UInt64 COMMENT 'Number of GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsRedirects` UInt64 COMMENT 'Number of redirects in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteMicroseconds` UInt64 COMMENT 'Time of POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsCount` UInt64 COMMENT 'Number of POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsRedirects` UInt64 COMMENT 'Number of redirects in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_S3DeleteObjects` UInt64 COMMENT 'Number of S3 API DeleteObject(s) calls.', `ProfileEvent_S3CopyObject` UInt64 COMMENT 'Number of S3 API CopyObject calls.', `ProfileEvent_S3ListObjects` UInt64 COMMENT 'Number of S3 API ListObjects calls.', `ProfileEvent_S3HeadObject` UInt64 COMMENT 'Number of S3 API HeadObject calls.', `ProfileEvent_S3GetObjectAttributes` UInt64 COMMENT 'Number of S3 API GetObjectAttributes calls.', `ProfileEvent_S3CreateMultipartUpload` UInt64 COMMENT 'Number of S3 API CreateMultipartUpload calls.', `ProfileEvent_S3UploadPartCopy` UInt64 COMMENT 'Number of S3 API UploadPartCopy calls.', `ProfileEvent_S3UploadPart` UInt64 COMMENT 'Number of S3 API UploadPart calls.', `ProfileEvent_S3AbortMultipartUpload` UInt64 COMMENT 'Number of S3 API AbortMultipartUpload calls.', `ProfileEvent_S3CompleteMultipartUpload` UInt64 COMMENT 'Number of S3 API CompleteMultipartUpload calls.', `ProfileEvent_S3PutObject` UInt64 COMMENT 'Number of S3 API PutObject calls.', `ProfileEvent_S3GetObject` UInt64 COMMENT 'Number of S3 API GetObject calls.', `ProfileEvent_DiskS3DeleteObjects` UInt64 COMMENT 'Number of DiskS3 API DeleteObject(s) calls.', `ProfileEvent_DiskS3CopyObject` UInt64 COMMENT 'Number of DiskS3 API CopyObject calls.', `ProfileEvent_DiskS3ListObjects` UInt64 COMMENT 'Number of DiskS3 API ListObjects calls.', `ProfileEvent_DiskS3HeadObject` UInt64 COMMENT 'Number of DiskS3 API HeadObject calls.', `ProfileEvent_DiskS3GetObjectAttributes` UInt64 COMMENT 'Number of DiskS3 API GetObjectAttributes calls.', `ProfileEvent_DiskS3CreateMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API CreateMultipartUpload calls.', `ProfileEvent_DiskS3UploadPartCopy` UInt64 COMMENT 'Number of DiskS3 API UploadPartCopy calls.', `ProfileEvent_DiskS3UploadPart` UInt64 COMMENT 'Number of DiskS3 API UploadPart calls.', `ProfileEvent_DiskS3AbortMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API AbortMultipartUpload calls.', `ProfileEvent_DiskS3CompleteMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API CompleteMultipartUpload calls.', `ProfileEvent_DiskS3PutObject` UInt64 COMMENT 'Number of DiskS3 API PutObject calls.', `ProfileEvent_DiskS3GetObject` UInt64 COMMENT 'Number of DiskS3 API GetObject calls.', `ProfileEvent_DiskPlainRewritableAzureDirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for AzureObjectStorage.', `ProfileEvent_DiskPlainRewritableAzureDirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for AzureObjectStorage.', `ProfileEvent_DiskPlainRewritableLocalDirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for LocalObjectStorage.', `ProfileEvent_DiskPlainRewritableLocalDirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for LocalObjectStorage.', `ProfileEvent_DiskPlainRewritableS3DirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for S3ObjectStorage.', `ProfileEvent_DiskPlainRewritableS3DirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for S3ObjectStorage.', `ProfileEvent_S3Clients` UInt64 COMMENT 'Number of created S3 clients.', `ProfileEvent_TinyS3Clients` UInt64 COMMENT 'Number of S3 clients copies which reuse an existing auth provider from another client.', `ProfileEvent_EngineFileLikeReadFiles` UInt64 COMMENT 'Number of files read in table engines working with files (like File/S3/URL/HDFS).', `ProfileEvent_ReadBufferFromS3Microseconds` UInt64 COMMENT 'Time spent on reading from S3.', `ProfileEvent_ReadBufferFromS3InitMicroseconds` UInt64 COMMENT 'Time spent initializing connection to S3.', `ProfileEvent_ReadBufferFromS3Bytes` UInt64 COMMENT 'Bytes read from S3.', `ProfileEvent_ReadBufferFromS3RequestsErrors` UInt64 COMMENT 'Number of exceptions while reading from S3.', `ProfileEvent_WriteBufferFromS3Microseconds` UInt64 COMMENT 'Time spent on writing to S3.', `ProfileEvent_WriteBufferFromS3Bytes` UInt64 COMMENT 'Bytes written to S3.', `ProfileEvent_WriteBufferFromS3RequestsErrors` UInt64 COMMENT 'Number of exceptions while writing to S3.', `ProfileEvent_WriteBufferFromS3WaitInflightLimitMicroseconds` UInt64 COMMENT 'Time spent on waiting while some of the current requests are done when its number reached the limit defined by s3_max_inflight_parts_for_one_file.', `ProfileEvent_QueryMemoryLimitExceeded` UInt64 COMMENT 'Number of times when memory limit exceeded for query.', `ProfileEvent_AzureGetObject` UInt64 COMMENT 'Number of Azure API GetObject calls.', `ProfileEvent_AzureUpload` UInt64 COMMENT 'Number of Azure blob storage API Upload calls', `ProfileEvent_AzureStageBlock` UInt64 COMMENT 'Number of Azure blob storage API StageBlock calls', `ProfileEvent_AzureCommitBlockList` UInt64 COMMENT 'Number of Azure blob storage API CommitBlockList calls', `ProfileEvent_AzureCopyObject` UInt64 COMMENT 'Number of Azure blob storage API CopyObject calls', `ProfileEvent_AzureDeleteObjects` UInt64 COMMENT 'Number of Azure blob storage API DeleteObject(s) calls.', `ProfileEvent_AzureListObjects` UInt64 COMMENT 'Number of Azure blob storage API ListObjects calls.', `ProfileEvent_AzureGetProperties` UInt64 COMMENT 'Number of Azure blob storage API GetProperties calls.', `ProfileEvent_AzureCreateContainer` UInt64 COMMENT 'Number of Azure blob storage API CreateContainer calls.', `ProfileEvent_DiskAzureGetObject` UInt64 COMMENT 'Number of Disk Azure API GetObject calls.', `ProfileEvent_DiskAzureUpload` UInt64 COMMENT 'Number of Disk Azure blob storage API Upload calls', `ProfileEvent_DiskAzureStageBlock` UInt64 COMMENT 'Number of Disk Azure blob storage API StageBlock calls', `ProfileEvent_DiskAzureCommitBlockList` UInt64 COMMENT 'Number of Disk Azure blob storage API CommitBlockList calls', `ProfileEvent_DiskAzureCopyObject` UInt64 COMMENT 'Number of Disk Azure blob storage API CopyObject calls', `ProfileEvent_DiskAzureListObjects` UInt64 COMMENT 'Number of Disk Azure blob storage API ListObjects calls.', `ProfileEvent_DiskAzureDeleteObjects` UInt64 COMMENT 'Number of Azure blob storage API DeleteObject(s) calls.', `ProfileEvent_DiskAzureGetProperties` UInt64 COMMENT 'Number of Disk Azure blob storage API GetProperties calls.', `ProfileEvent_DiskAzureCreateContainer` UInt64 COMMENT 'Number of Disk Azure blob storage API CreateContainer calls.', `ProfileEvent_ReadBufferFromAzureMicroseconds` UInt64 COMMENT 'Time spent on reading from Azure.', `ProfileEvent_ReadBufferFromAzureInitMicroseconds` UInt64 COMMENT 'Time spent initializing connection to Azure.', `ProfileEvent_ReadBufferFromAzureBytes` UInt64 COMMENT 'Bytes read from Azure.', `ProfileEvent_ReadBufferFromAzureRequestsErrors` UInt64 COMMENT 'Number of exceptions while reading from Azure', `ProfileEvent_CachedReadBufferReadFromCacheHits` UInt64 COMMENT 'Number of times the read from filesystem cache hit the cache.', `ProfileEvent_CachedReadBufferReadFromCacheMisses` UInt64 COMMENT 'Number of times the read from filesystem cache miss the cache.', `ProfileEvent_CachedReadBufferReadFromSourceMicroseconds` UInt64 COMMENT 'Time reading from filesystem cache source (from remote filesystem, etc)', `ProfileEvent_CachedReadBufferReadFromCacheMicroseconds` UInt64 COMMENT 'Time reading from filesystem cache', `ProfileEvent_CachedReadBufferReadFromSourceBytes` UInt64 COMMENT 'Bytes read from filesystem cache source (from remote fs, etc)', `ProfileEvent_CachedReadBufferReadFromCacheBytes` UInt64 COMMENT 'Bytes read from filesystem cache', `ProfileEvent_CachedReadBufferPredownloadedBytes` UInt64 COMMENT 'Bytes read from filesystem cache source. Cache segments are read from left to right as a whole, it might be that we need to predownload some part of the segment irrelevant for the current task just to get to the needed data', `ProfileEvent_CachedReadBufferCacheWriteBytes` UInt64 COMMENT 'Bytes written from source (remote fs, etc) to filesystem cache', `ProfileEvent_CachedReadBufferCacheWriteMicroseconds` UInt64 COMMENT 'Time spent writing data into filesystem cache', `ProfileEvent_CachedReadBufferCreateBufferMicroseconds` UInt64 COMMENT 'Prepare buffer time', `ProfileEvent_CachedWriteBufferCacheWriteBytes` UInt64 COMMENT 'Bytes written from source (remote fs, etc) to filesystem cache', `ProfileEvent_CachedWriteBufferCacheWriteMicroseconds` UInt64 COMMENT 'Time spent writing data into filesystem cache', `ProfileEvent_FilesystemCacheLoadMetadataMicroseconds` UInt64 COMMENT 'Time spent loading filesystem cache metadata', `ProfileEvent_FilesystemCacheEvictedBytes` UInt64 COMMENT 'Number of bytes evicted from filesystem cache', `ProfileEvent_FilesystemCacheEvictedFileSegments` UInt64 COMMENT 'Number of file segments evicted from filesystem cache', `ProfileEvent_FilesystemCacheBackgroundDownloadQueuePush` UInt64 COMMENT 'Number of file segments sent for background download in filesystem cache', `ProfileEvent_FilesystemCacheEvictionSkippedFileSegments` UInt64 COMMENT 'Number of file segments skipped for eviction because of being in unreleasable state', `ProfileEvent_FilesystemCacheEvictionSkippedEvictingFileSegments` UInt64 COMMENT 'Number of file segments skipped for eviction because of being in evicting state', `ProfileEvent_FilesystemCacheEvictionTries` UInt64 COMMENT 'Number of filesystem cache eviction attempts', `ProfileEvent_FilesystemCacheLockKeyMicroseconds` UInt64 COMMENT 'Lock cache key time', `ProfileEvent_FilesystemCacheLockMetadataMicroseconds` UInt64 COMMENT 'Lock filesystem cache metadata time', `ProfileEvent_FilesystemCacheLockCacheMicroseconds` UInt64 COMMENT 'Lock filesystem cache time', `ProfileEvent_FilesystemCacheReserveMicroseconds` UInt64 COMMENT 'Filesystem cache space reservation time', `ProfileEvent_FilesystemCacheEvictMicroseconds` UInt64 COMMENT 'Filesystem cache eviction time', `ProfileEvent_FilesystemCacheGetOrSetMicroseconds` UInt64 COMMENT 'Filesystem cache getOrSet() time', `ProfileEvent_FilesystemCacheGetMicroseconds` UInt64 COMMENT 'Filesystem cache get() time', `ProfileEvent_FileSegmentWaitMicroseconds` UInt64 COMMENT 'Wait on DOWNLOADING state', `ProfileEvent_FileSegmentCompleteMicroseconds` UInt64 COMMENT 'Duration of FileSegment::complete() in filesystem cache', `ProfileEvent_FileSegmentLockMicroseconds` UInt64 COMMENT 'Lock file segment time', `ProfileEvent_FileSegmentWriteMicroseconds` UInt64 COMMENT 'File segment write() time', `ProfileEvent_FileSegmentUseMicroseconds` UInt64 COMMENT 'File segment use() time', `ProfileEvent_FileSegmentRemoveMicroseconds` UInt64 COMMENT 'File segment remove() time', `ProfileEvent_FileSegmentHolderCompleteMicroseconds` UInt64 COMMENT 'File segments holder complete() time', `ProfileEvent_FileSegmentFailToIncreasePriority` UInt64 COMMENT 'Number of times the priority was not increased due to a high contention on the cache lock', `ProfileEvent_FilesystemCacheFailToReserveSpaceBecauseOfLockContention` UInt64 COMMENT 'Number of times space reservation was skipped due to a high contention on the cache lock', `ProfileEvent_FilesystemCacheFailToReserveSpaceBecauseOfCacheResize` UInt64 COMMENT 'Number of times space reservation was skipped due to the cache is being resized', `ProfileEvent_FilesystemCacheHoldFileSegments` UInt64 COMMENT 'Filesystem cache file segments count, which were hold', `ProfileEvent_FilesystemCacheUnusedHoldFileSegments` UInt64 COMMENT 'Filesystem cache file segments count, which were hold, but not used (because of seek or LIMIT n, etc)', `ProfileEvent_FilesystemCacheFreeSpaceKeepingThreadRun` UInt64 COMMENT 'Number of times background thread executed free space keeping job', `ProfileEvent_FilesystemCacheFreeSpaceKeepingThreadWorkMilliseconds` UInt64 COMMENT 'Time for which background thread executed free space keeping job', `ProfileEvent_RemoteFSSeeks` UInt64 COMMENT 'Total number of seeks for async buffer', `ProfileEvent_RemoteFSPrefetches` UInt64 COMMENT 'Number of prefetches made with asynchronous reading from remote filesystem', `ProfileEvent_RemoteFSCancelledPrefetches` UInt64 COMMENT 'Number of cancelled prefecthes (because of seek)', `ProfileEvent_RemoteFSUnusedPrefetches` UInt64 COMMENT 'Number of prefetches pending at buffer destruction', `ProfileEvent_RemoteFSPrefetchedReads` UInt64 COMMENT 'Number of reads from prefecthed buffer', `ProfileEvent_RemoteFSPrefetchedBytes` UInt64 COMMENT 'Number of bytes from prefecthed buffer', `ProfileEvent_RemoteFSUnprefetchedReads` UInt64 COMMENT 'Number of reads from unprefetched buffer', `ProfileEvent_RemoteFSUnprefetchedBytes` UInt64 COMMENT 'Number of bytes from unprefetched buffer', `ProfileEvent_RemoteFSLazySeeks` UInt64 COMMENT 'Number of lazy seeks', `ProfileEvent_RemoteFSSeeksWithReset` UInt64 COMMENT 'Number of seeks which lead to a new connection', `ProfileEvent_RemoteFSBuffers` UInt64 COMMENT 'Number of buffers created for asynchronous reading from remote filesystem', `ProfileEvent_MergeTreePrefetchedReadPoolInit` UInt64 COMMENT 'Time spent preparing tasks in MergeTreePrefetchedReadPool', `ProfileEvent_WaitPrefetchTaskMicroseconds` UInt64 COMMENT 'Time spend waiting for prefetched reader', `ProfileEvent_ThreadpoolReaderTaskMicroseconds` UInt64 COMMENT 'Time spent getting the data in asynchronous reading', `ProfileEvent_ThreadpoolReaderPrepareMicroseconds` UInt64 COMMENT 'Time spent on preparation (e.g. call to reader seek() method)', `ProfileEvent_ThreadpoolReaderReadBytes` UInt64 COMMENT 'Bytes read from a threadpool task in asynchronous reading', `ProfileEvent_ThreadpoolReaderSubmit` UInt64 COMMENT 'Bytes read from a threadpool task in asynchronous reading', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronously` UInt64 COMMENT 'How many times we haven\'t scheduled a task on the thread pool and read synchronously instead', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronouslyBytes` UInt64 COMMENT 'How many bytes were read synchronously', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronouslyMicroseconds` UInt64 COMMENT 'How much time we spent reading synchronously', `ProfileEvent_ThreadpoolReaderSubmitLookupInCacheMicroseconds` UInt64 COMMENT 'How much time we spent checking if content is cached', `ProfileEvent_AsynchronousReaderIgnoredBytes` UInt64 COMMENT 'Number of bytes ignored during asynchronous reading', `ProfileEvent_FileSegmentWaitReadBufferMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend waiting for internal read buffer (includes cache waiting)', `ProfileEvent_FileSegmentReadMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend reading from file', `ProfileEvent_FileSegmentCacheWriteMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend writing data to cache', `ProfileEvent_FileSegmentPredownloadMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spent pre-downloading data to cache (pre-downloading - finishing file segment download (after someone who failed to do that) up to the point current thread was requested to do)', `ProfileEvent_FileSegmentUsedBytes` UInt64 COMMENT 'Metric per file segment. How many bytes were actually used from current file segment', `ProfileEvent_ReadBufferSeekCancelConnection` UInt64 COMMENT 'Number of seeks which lead to new connection (s3, http)', `ProfileEvent_SleepFunctionCalls` UInt64 COMMENT 'Number of times a sleep function (sleep, sleepEachRow) has been called.', `ProfileEvent_SleepFunctionMicroseconds` UInt64 COMMENT 'Time set to sleep in a sleep function (sleep, sleepEachRow).', `ProfileEvent_SleepFunctionElapsedMicroseconds` UInt64 COMMENT 'Time spent sleeping in a sleep function (sleep, sleepEachRow).', `ProfileEvent_ThreadPoolReaderPageCacheHit` UInt64 COMMENT 'Number of times the read inside ThreadPoolReader was done from the page cache.', `ProfileEvent_ThreadPoolReaderPageCacheHitBytes` UInt64 COMMENT 'Number of bytes read inside ThreadPoolReader when it was done from the page cache.', `ProfileEvent_ThreadPoolReaderPageCacheHitElapsedMicroseconds` UInt64 COMMENT 'Time spent reading data from page cache in ThreadPoolReader.', `ProfileEvent_ThreadPoolReaderPageCacheMiss` UInt64 COMMENT 'Number of times the read inside ThreadPoolReader was not done from page cache and was hand off to thread pool.', `ProfileEvent_ThreadPoolReaderPageCacheMissBytes` UInt64 COMMENT 'Number of bytes read inside ThreadPoolReader when read was not done from page cache and was hand off to thread pool.', `ProfileEvent_ThreadPoolReaderPageCacheMissElapsedMicroseconds` UInt64 COMMENT 'Time spent reading data inside the asynchronous job in ThreadPoolReader - when read was not done from the page cache.', `ProfileEvent_AsynchronousReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for asynchronous reads in asynchronous local read.', `ProfileEvent_SynchronousReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for synchronous reads in asynchronous local read.', `ProfileEvent_AsynchronousRemoteReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for asynchronous remote reads.', `ProfileEvent_SynchronousRemoteReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for synchronous remote reads.', `ProfileEvent_ExternalDataSourceLocalCacheReadBytes` UInt64 COMMENT 'Bytes read from local cache buffer in RemoteReadBufferCache', `ProfileEvent_MainConfigLoads` UInt64 COMMENT 'Number of times the main configuration was reloaded.', `ProfileEvent_AggregationPreallocatedElementsInHashTables` UInt64 COMMENT 'How many elements were preallocated in hash tables for aggregation.', `ProfileEvent_AggregationHashTablesInitializedAsTwoLevel` UInt64 COMMENT 'How many hash tables were inited as two-level for aggregation.', `ProfileEvent_AggregationOptimizedEqualRangesOfKeys` UInt64 COMMENT 'For how many blocks optimization of equal ranges of keys was applied', `ProfileEvent_HashJoinPreallocatedElementsInHashTables` UInt64 COMMENT 'How many elements were preallocated in hash tables for hash join.', `ProfileEvent_MetadataFromKeeperCacheHit` UInt64 COMMENT 'Number of times an object storage metadata request was answered from cache without making request to Keeper', `ProfileEvent_MetadataFromKeeperCacheMiss` UInt64 COMMENT 'Number of times an object storage metadata request had to be answered from Keeper', `ProfileEvent_MetadataFromKeeperCacheUpdateMicroseconds` UInt64 COMMENT 'Total time spent in updating the cache including waiting for responses from Keeper', `ProfileEvent_MetadataFromKeeperUpdateCacheOneLevel` UInt64 COMMENT 'Number of times a cache update for one level of directory tree was done', `ProfileEvent_MetadataFromKeeperTransactionCommit` UInt64 COMMENT 'Number of times metadata transaction commit was attempted', `ProfileEvent_MetadataFromKeeperTransactionCommitRetry` UInt64 COMMENT 'Number of times metadata transaction commit was retried', `ProfileEvent_MetadataFromKeeperCleanupTransactionCommit` UInt64 COMMENT 'Number of times metadata transaction commit for deleted objects cleanup was attempted', `ProfileEvent_MetadataFromKeeperCleanupTransactionCommitRetry` UInt64 COMMENT 'Number of times metadata transaction commit for deleted objects cleanup was retried', `ProfileEvent_MetadataFromKeeperOperations` UInt64 COMMENT 'Number of times a request was made to Keeper', `ProfileEvent_MetadataFromKeeperIndividualOperations` UInt64 COMMENT 'Number of paths read or written by single or multi requests to Keeper', `ProfileEvent_MetadataFromKeeperReconnects` UInt64 COMMENT 'Number of times a reconnect to Keeper was done', `ProfileEvent_MetadataFromKeeperBackgroundCleanupObjects` UInt64 COMMENT 'Number of times a old deleted object clean up was performed by background task', `ProfileEvent_MetadataFromKeeperBackgroundCleanupTransactions` UInt64 COMMENT 'Number of times old transaction idempotency token was cleaned up by background task', `ProfileEvent_MetadataFromKeeperBackgroundCleanupErrors` UInt64 COMMENT 'Number of times an error was encountered in background cleanup task', `ProfileEvent_SharedMergeTreeMetadataCacheHintLoadedFromCache` UInt64 COMMENT 'Number of times metadata cache hint was found without going to Keeper', `ProfileEvent_KafkaRebalanceRevocations` UInt64 COMMENT 'Number of partition revocations (the first stage of consumer group rebalance)', `ProfileEvent_KafkaRebalanceAssignments` UInt64 COMMENT 'Number of partition assignments (the final stage of consumer group rebalance)', `ProfileEvent_KafkaRebalanceErrors` UInt64 COMMENT 'Number of failed consumer group rebalances', `ProfileEvent_KafkaMessagesPolled` UInt64 COMMENT 'Number of Kafka messages polled from librdkafka to ClickHouse', `ProfileEvent_KafkaMessagesRead` UInt64 COMMENT 'Number of Kafka messages already processed by ClickHouse', `ProfileEvent_KafkaMessagesFailed` UInt64 COMMENT 'Number of Kafka messages ClickHouse failed to parse', `ProfileEvent_KafkaRowsRead` UInt64 COMMENT 'Number of rows parsed from Kafka messages', `ProfileEvent_KafkaRowsRejected` UInt64 COMMENT 'Number of parsed rows which were later rejected (due to rebalances / errors or similar reasons). Those rows will be consumed again after the rebalance.', `ProfileEvent_KafkaDirectReads` UInt64 COMMENT 'Number of direct selects from Kafka tables since server start', `ProfileEvent_KafkaBackgroundReads` UInt64 COMMENT 'Number of background reads populating materialized views from Kafka since server start', `ProfileEvent_KafkaCommits` UInt64 COMMENT 'Number of successful commits of consumed offsets to Kafka (normally should be the same as KafkaBackgroundReads)', `ProfileEvent_KafkaCommitFailures` UInt64 COMMENT 'Number of failed commits of consumed offsets to Kafka (usually is a sign of some data duplication)', `ProfileEvent_KafkaConsumerErrors` UInt64 COMMENT 'Number of errors reported by librdkafka during polls', `ProfileEvent_KafkaWrites` UInt64 COMMENT 'Number of writes (inserts) to Kafka tables ', `ProfileEvent_KafkaRowsWritten` UInt64 COMMENT 'Number of rows inserted into Kafka tables', `ProfileEvent_KafkaProducerFlushes` UInt64 COMMENT 'Number of explicit flushes to Kafka producer', `ProfileEvent_KafkaMessagesProduced` UInt64 COMMENT 'Number of messages produced to Kafka', `ProfileEvent_KafkaProducerErrors` UInt64 COMMENT 'Number of errors during producing the messages to Kafka', `ProfileEvent_ScalarSubqueriesGlobalCacheHit` UInt64 COMMENT 'Number of times a read from a scalar subquery was done using the global cache', `ProfileEvent_ScalarSubqueriesLocalCacheHit` UInt64 COMMENT 'Number of times a read from a scalar subquery was done using the local cache', `ProfileEvent_ScalarSubqueriesCacheMiss` UInt64 COMMENT 'Number of times a read from a scalar subquery was not cached and had to be calculated completely', `ProfileEvent_SchemaInferenceCacheHits` UInt64 COMMENT 'Number of times the requested source is found in schema cache', `ProfileEvent_SchemaInferenceCacheSchemaHits` UInt64 COMMENT 'Number of times the schema is found in schema cache during schema inference', `ProfileEvent_SchemaInferenceCacheNumRowsHits` UInt64 COMMENT 'Number of times the number of rows is found in schema cache during count from files', `ProfileEvent_SchemaInferenceCacheMisses` UInt64 COMMENT 'Number of times the requested source is not in schema cache', `ProfileEvent_SchemaInferenceCacheSchemaMisses` UInt64 COMMENT 'Number of times the requested source is in cache but the schema is not in cache during schema inference', `ProfileEvent_SchemaInferenceCacheNumRowsMisses` UInt64 COMMENT 'Number of times the requested source is in cache but the number of rows is not in cache while count from files', `ProfileEvent_SchemaInferenceCacheEvictions` UInt64 COMMENT 'Number of times a schema from cache was evicted due to overflow', `ProfileEvent_SchemaInferenceCacheInvalidations` UInt64 COMMENT 'Number of times a schema in cache became invalid due to changes in data', `ProfileEvent_KeeperPacketsSent` UInt64 COMMENT 'Packets sent by keeper server', `ProfileEvent_KeeperPacketsReceived` UInt64 COMMENT 'Packets received by keeper server', `ProfileEvent_KeeperRequestTotal` UInt64 COMMENT 'Total requests number on keeper server', `ProfileEvent_KeeperLatency` UInt64 COMMENT 'Keeper latency', `ProfileEvent_KeeperTotalElapsedMicroseconds` UInt64 COMMENT 'Keeper total latency for a single request', `ProfileEvent_KeeperProcessElapsedMicroseconds` UInt64 COMMENT 'Keeper commit latency for a single request', `ProfileEvent_KeeperPreprocessElapsedMicroseconds` UInt64 COMMENT 'Keeper preprocessing latency for a single reuquest', `ProfileEvent_KeeperStorageLockWaitMicroseconds` UInt64 COMMENT 'Time spent waiting for acquiring Keeper storage lock', `ProfileEvent_KeeperCommitWaitElapsedMicroseconds` UInt64 COMMENT 'Time spent waiting for certain log to be committed', `ProfileEvent_KeeperBatchMaxCount` UInt64 COMMENT 'Number of times the size of batch was limited by the amount', `ProfileEvent_KeeperBatchMaxTotalSize` UInt64 COMMENT 'Number of times the size of batch was limited by the total bytes size', `ProfileEvent_KeeperCommits` UInt64 COMMENT 'Number of successful commits', `ProfileEvent_KeeperCommitsFailed` UInt64 COMMENT 'Number of failed commits', `ProfileEvent_KeeperSnapshotCreations` UInt64 COMMENT 'Number of snapshots creations', `ProfileEvent_KeeperSnapshotCreationsFailed` UInt64 COMMENT 'Number of failed snapshot creations', `ProfileEvent_KeeperSnapshotApplys` UInt64 COMMENT 'Number of snapshot applying', `ProfileEvent_KeeperSnapshotApplysFailed` UInt64 COMMENT 'Number of failed snapshot applying', `ProfileEvent_KeeperReadSnapshot` UInt64 COMMENT 'Number of snapshot read(serialization)', `ProfileEvent_KeeperSaveSnapshot` UInt64 COMMENT 'Number of snapshot save', `ProfileEvent_KeeperCreateRequest` UInt64 COMMENT 'Number of create requests', `ProfileEvent_KeeperRemoveRequest` UInt64 COMMENT 'Number of remove requests', `ProfileEvent_KeeperSetRequest` UInt64 COMMENT 'Number of set requests', `ProfileEvent_KeeperReconfigRequest` UInt64 COMMENT 'Number of reconfig requests', `ProfileEvent_KeeperCheckRequest` UInt64 COMMENT 'Number of check requests', `ProfileEvent_KeeperMultiRequest` UInt64 COMMENT 'Number of multi requests', `ProfileEvent_KeeperMultiReadRequest` UInt64 COMMENT 'Number of multi read requests', `ProfileEvent_KeeperGetRequest` UInt64 COMMENT 'Number of get requests', `ProfileEvent_KeeperListRequest` UInt64 COMMENT 'Number of list requests', `ProfileEvent_KeeperExistsRequest` UInt64 COMMENT 'Number of exists requests', `ProfileEvent_OverflowBreak` UInt64 COMMENT 'Number of times, data processing was cancelled by query complexity limitation with setting \'*_overflow_mode\' = \'break\' and the result is incomplete.', `ProfileEvent_OverflowThrow` UInt64 COMMENT 'Number of times, data processing was cancelled by query complexity limitation with setting \'*_overflow_mode\' = \'throw\' and exception was thrown.', `ProfileEvent_OverflowAny` UInt64 COMMENT 'Number of times approximate GROUP BY was in effect: when aggregation was performed only on top of first \'max_rows_to_group_by\' unique keys and other keys were ignored due to \'group_by_overflow_mode\' = \'any\'.', `ProfileEvent_S3QueueSetFileProcessingMicroseconds` UInt64 COMMENT 'Time spent to set file as processing', `ProfileEvent_S3QueueSetFileProcessedMicroseconds` UInt64 COMMENT 'Time spent to set file as processed', `ProfileEvent_S3QueueSetFileFailedMicroseconds` UInt64 COMMENT 'Time spent to set file as failed', `ProfileEvent_ObjectStorageQueueFailedFiles` UInt64 COMMENT 'Number of files which failed to be processed', `ProfileEvent_ObjectStorageQueueProcessedFiles` UInt64 COMMENT 'Number of files which were processed', `ProfileEvent_ObjectStorageQueueCleanupMaxSetSizeOrTTLMicroseconds` UInt64 COMMENT 'Time spent to set file as failed', `ProfileEvent_ObjectStorageQueuePullMicroseconds` UInt64 COMMENT 'Time spent to read file data', `ProfileEvent_ObjectStorageQueueLockLocalFileStatusesMicroseconds` UInt64 COMMENT 'Time spent to lock local file statuses', `ProfileEvent_ObjectStorageQueueFailedToBatchSetProcessing` UInt64 COMMENT 'Number of times batched set processing request failed', `ProfileEvent_ObjectStorageQueueTrySetProcessingRequests` UInt64 COMMENT 'The number of times we tried to make set processing request', `ProfileEvent_ObjectStorageQueueTrySetProcessingSucceeded` UInt64 COMMENT 'The number of times we successfully set file as processing', `ProfileEvent_ObjectStorageQueueTrySetProcessingFailed` UInt64 COMMENT 'The number of times we unsuccessfully set file as processing', `ProfileEvent_ObjectStorageQueueListedFiles` UInt64 COMMENT 'Number of listed files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueFilteredFiles` UInt64 COMMENT 'Number of filtered files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueReadFiles` UInt64 COMMENT 'Number of read files (not equal to the number of actually inserted files)', `ProfileEvent_ObjectStorageQueueReadRows` UInt64 COMMENT 'Number of read rows (not equal to the number of actually inserted rows)', `ProfileEvent_ObjectStorageQueueReadBytes` UInt64 COMMENT 'Number of read bytes (not equal to the number of actually inserted bytes)', `ProfileEvent_ObjectStorageQueueExceptionsDuringRead` UInt64 COMMENT 'Number of exceptions during read in S3(Azure)Queue', `ProfileEvent_ObjectStorageQueueExceptionsDuringInsert` UInt64 COMMENT 'Number of exceptions during insert in S3(Azure)Queue', `ProfileEvent_ObjectStorageQueueRemovedObjects` UInt64 COMMENT 'Number of objects removed as part of after_processing = delete', `ProfileEvent_ObjectStorageQueueInsertIterations` UInt64 COMMENT 'Number of insert iterations', `ProfileEvent_ObjectStorageQueueCommitRequests` UInt64 COMMENT 'Number of keeper requests to commit files as either failed or processed', `ProfileEvent_ObjectStorageQueueSuccessfulCommits` UInt64 COMMENT 'Number of successful keeper commits', `ProfileEvent_ObjectStorageQueueUnsuccessfulCommits` UInt64 COMMENT 'Number of unsuccessful keeper commits', `ProfileEvent_ObjectStorageQueueCancelledFiles` UInt64 COMMENT 'Number cancelled files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueProcessedRows` UInt64 COMMENT 'Number of processed rows in StorageS3(Azure)Queue', `ProfileEvent_ServerStartupMilliseconds` UInt64 COMMENT 'Time elapsed from starting server to listening to sockets in milliseconds', `ProfileEvent_IOUringSQEsSubmitted` UInt64 COMMENT 'Total number of io_uring SQEs submitted', `ProfileEvent_IOUringSQEsResubmitsAsync` UInt64 COMMENT 'Total number of asynchronous io_uring SQE resubmits performed', `ProfileEvent_IOUringSQEsResubmitsSync` UInt64 COMMENT 'Total number of synchronous io_uring SQE resubmits performed', `ProfileEvent_IOUringCQEsCompleted` UInt64 COMMENT 'Total number of successfully completed io_uring CQEs', `ProfileEvent_IOUringCQEsFailed` UInt64 COMMENT 'Total number of completed io_uring CQEs with failures', `ProfileEvent_BackupsOpenedForRead` UInt64 COMMENT 'Number of backups opened for reading', `ProfileEvent_BackupsOpenedForWrite` UInt64 COMMENT 'Number of backups opened for writing', `ProfileEvent_BackupsOpenedForUnlock` UInt64 COMMENT 'Number of backups opened for unlocking', `ProfileEvent_BackupReadMetadataMicroseconds` UInt64 COMMENT 'Time spent reading backup metadata from .backup file', `ProfileEvent_BackupWriteMetadataMicroseconds` UInt64 COMMENT 'Time spent writing backup metadata to .backup file', `ProfileEvent_BackupEntriesCollectorMicroseconds` UInt64 COMMENT 'Time spent making backup entries', `ProfileEvent_BackupEntriesCollectorForTablesDataMicroseconds` UInt64 COMMENT 'Time spent making backup entries for tables data', `ProfileEvent_BackupEntriesCollectorRunPostTasksMicroseconds` UInt64 COMMENT 'Time spent running post tasks after making backup entries', `ProfileEvent_BackupPreparingFileInfosMicroseconds` UInt64 COMMENT 'Time spent preparing file infos for backup entries', `ProfileEvent_BackupReadLocalFilesToCalculateChecksums` UInt64 COMMENT 'Number of files read locally to calculate checksums for backup entries', `ProfileEvent_BackupReadLocalBytesToCalculateChecksums` UInt64 COMMENT 'Total size of files read locally to calculate checksums for backup entries', `ProfileEvent_BackupReadRemoteFilesToCalculateChecksums` UInt64 COMMENT 'Number of files read from remote disks to calculate checksums for backup entries', `ProfileEvent_BackupReadRemoteBytesToCalculateChecksums` UInt64 COMMENT 'Total size of files read from remote disks to calculate checksums for backup entries', `ProfileEvent_BackupLockFileReads` UInt64 COMMENT 'How many times the \'.lock\' file was read while making backup', `ProfileEvent_RestorePartsSkippedFiles` UInt64 COMMENT 'Number of files skipped while restoring parts', `ProfileEvent_RestorePartsSkippedBytes` UInt64 COMMENT 'Total size of files skipped while restoring parts', `ProfileEvent_ReadTaskRequestsReceived` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the initiator server side.', `ProfileEvent_MergeTreeReadTaskRequestsReceived` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the initiator server side.', `ProfileEvent_ReadTaskRequestsSent` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `ProfileEvent_MergeTreeReadTaskRequestsSent` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergeTreeAllRangesAnnouncementsSent` UInt64 COMMENT 'The number of announcements sent from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_ReadTaskRequestsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `ProfileEvent_MergeTreeReadTaskRequestsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergeTreeAllRangesAnnouncementsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in sending the announcement from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergerMutatorsGetPartsForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to take data parts snapshot to build ranges from them.', `ProfileEvent_MergerMutatorPrepareRangesForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to prepare parts ranges which can be merged according to merge predicate.', `ProfileEvent_MergerMutatorSelectPartsForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to select parts from ranges which can be merged.', `ProfileEvent_MergerMutatorRangesForMergeCount` UInt64 COMMENT 'Amount of candidate ranges for merge', `ProfileEvent_MergerMutatorPartsInRangesForMergeCount` UInt64 COMMENT 'Amount of candidate parts for merge', `ProfileEvent_MergerMutatorSelectRangePartsCount` UInt64 COMMENT 'Amount of parts in selected range for merge', `ProfileEvent_ConnectionPoolIsFullMicroseconds` UInt64 COMMENT 'Total time spent waiting for a slot in connection pool.', `ProfileEvent_AsyncLoaderWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting for async loader jobs.', `ProfileEvent_DistrCacheServerSwitches` UInt64 COMMENT 'Distributed Cache read buffer event. Number of server switches between distributed cache servers in read/write-through cache', `ProfileEvent_DistrCacheReadMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent reading from distributed cache', `ProfileEvent_DistrCacheFallbackReadMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spend reading from fallback buffer instead of distributed cache', `ProfileEvent_DistrCachePrecomputeRangesMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent to precompute read ranges', `ProfileEvent_DistrCacheNextImplMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spend in ReadBufferFromDistributedCache::nextImpl', `ProfileEvent_DistrCacheStartRangeMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent to start a new read range with distributed cache', `ProfileEvent_DistrCacheIgnoredBytesWhileWaitingProfileEvents` UInt64 COMMENT 'Distributed Cache read buffer event. Ignored bytes while waiting for profile events in distributed cache', `ProfileEvent_DistrCacheRangeChange` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we changed read range because of seek/last_position change', `ProfileEvent_DistrCacheGetResponseMicroseconds` UInt64 COMMENT 'Distributed Cache client event. Time spend to wait for response from distributed cache', `ProfileEvent_DistrCacheReadErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors during read', `ProfileEvent_DistrCacheMakeRequestErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors when making a request', `ProfileEvent_DistrCacheReceiveResponseErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors when receiving response a request', `ProfileEvent_DistrCachePackets` UInt64 COMMENT 'Distributed Cache client event. Total number of packets received from distributed cache', `ProfileEvent_DistrCachePacketsBytes` UInt64 COMMENT 'Distributed Cache client event. The number of bytes in Data packets which were not ignored', `ProfileEvent_DistrCacheUnusedPackets` UInt64 COMMENT 'Distributed Cache client event. Number of skipped unused packets from distributed cache', `ProfileEvent_DistrCacheUnusedPacketsBytes` UInt64 COMMENT 'Distributed Cache client event. The number of bytes in Data packets which were ignored', `ProfileEvent_DistrCacheUnusedPacketsBufferAllocations` UInt64 COMMENT 'Distributed Cache client event. The number of extra buffer allocations in case we could not reuse existing buffer', `ProfileEvent_DistrCacheLockRegistryMicroseconds` UInt64 COMMENT 'Distributed Cache registry event. Time spent to take DistributedCacheRegistry lock', `ProfileEvent_DistrCacheRegistryUpdateMicroseconds` UInt64 COMMENT 'Distributed Cache registry event. Time spent updating distributed cache registry', `ProfileEvent_DistrCacheRegistryUpdates` UInt64 COMMENT 'Distributed Cache registry event. Number of distributed cache registry updates', `ProfileEvent_DistrCacheHashRingRebuilds` UInt64 COMMENT 'Distributed Cache registry event. Number of distributed cache hash ring rebuilds', `ProfileEvent_DistrCacheReadBytesFromCache` UInt64 COMMENT 'Distributed Cache read buffer event. Bytes read from distributed cache', `ProfileEvent_DistrCacheReadBytesFromFallbackBuffer` UInt64 COMMENT 'Distributed Cache read buffer event. Bytes read from fallback buffer', `ProfileEvent_DistrCacheRangeResetBackward` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we reset read range because of seek/last_position change', `ProfileEvent_DistrCacheRangeResetForward` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we reset read range because of seek/last_position change', `ProfileEvent_DistrCacheOpenedConnections` UInt64 COMMENT 'Distributed Cache connection event. The number of open connections to distributed cache', `ProfileEvent_DistrCacheReusedConnections` UInt64 COMMENT 'Distributed Cache connection event. The number of reused connections to distributed cache', `ProfileEvent_DistrCacheOpenedConnectionsBypassingPool` UInt64 COMMENT 'Distributed Cache connection event. The number of open connections to distributed cache bypassing pool', `ProfileEvent_DistrCacheConnectMicroseconds` UInt64 COMMENT 'Distributed Cache connection event. The time spent to connect to distributed cache', `ProfileEvent_DistrCacheConnectAttempts` UInt64 COMMENT 'Distributed Cache connection event. The number of connection attempts to distributed cache', `ProfileEvent_DistrCacheGetClientMicroseconds` UInt64 COMMENT 'Distributed Cache connection event. Time spent getting client for distributed cache', `ProfileEvent_DistrCacheServerProcessRequestMicroseconds` UInt64 COMMENT 'Distributed Cache server event. Time spent processing request on DistributedCache server side', `ProfileEvent_DistrCacheServerStartRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of StartRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerContinueRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of ContinueRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerEndRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of EndRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerAckRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of AckRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerNewS3CachedClients` UInt64 COMMENT 'Distributed Cache server event. The number of new cached s3 clients', `ProfileEvent_DistrCacheServerReusedS3CachedClients` UInt64 COMMENT 'Distributed Cache server event. The number of reused cached s3 clients', `ProfileEvent_LogTest` UInt64 COMMENT 'Number of log messages with level Test', `ProfileEvent_LogTrace` UInt64 COMMENT 'Number of log messages with level Trace', `ProfileEvent_LogDebug` UInt64 COMMENT 'Number of log messages with level Debug', `ProfileEvent_LogInfo` UInt64 COMMENT 'Number of log messages with level Info', `ProfileEvent_LogWarning` UInt64 COMMENT 'Number of log messages with level Warning', `ProfileEvent_LogError` UInt64 COMMENT 'Number of log messages with level Error', `ProfileEvent_LogFatal` UInt64 COMMENT 'Number of log messages with level Fatal', `ProfileEvent_LoggerElapsedNanoseconds` UInt64 COMMENT 'Cumulative time spend in logging', `ProfileEvent_InterfaceHTTPSendBytes` UInt64 COMMENT 'Number of bytes sent through HTTP interfaces', `ProfileEvent_InterfaceHTTPReceiveBytes` UInt64 COMMENT 'Number of bytes received through HTTP interfaces', `ProfileEvent_InterfaceNativeSendBytes` UInt64 COMMENT 'Number of bytes sent through native interfaces', `ProfileEvent_InterfaceNativeReceiveBytes` UInt64 COMMENT 'Number of bytes received through native interfaces', `ProfileEvent_InterfacePrometheusSendBytes` UInt64 COMMENT 'Number of bytes sent through Prometheus interfaces', `ProfileEvent_InterfacePrometheusReceiveBytes` UInt64 COMMENT 'Number of bytes received through Prometheus interfaces', `ProfileEvent_InterfaceInterserverSendBytes` UInt64 COMMENT 'Number of bytes sent through interserver interfaces', `ProfileEvent_InterfaceInterserverReceiveBytes` UInt64 COMMENT 'Number of bytes received through interserver interfaces', `ProfileEvent_InterfaceMySQLSendBytes` UInt64 COMMENT 'Number of bytes sent through MySQL interfaces', `ProfileEvent_InterfaceMySQLReceiveBytes` UInt64 COMMENT 'Number of bytes received through MySQL interfaces', `ProfileEvent_InterfacePostgreSQLSendBytes` UInt64 COMMENT 'Number of bytes sent through PostgreSQL interfaces', `ProfileEvent_InterfacePostgreSQLReceiveBytes` UInt64 COMMENT 'Number of bytes received through PostgreSQL interfaces', `ProfileEvent_ParallelReplicasUsedCount` UInt64 COMMENT 'Number of replicas used to execute a query with task-based parallel replicas', `ProfileEvent_ParallelReplicasAvailableCount` UInt64 COMMENT 'Number of replicas available to execute a query with task-based parallel replicas', `ProfileEvent_ParallelReplicasUnavailableCount` UInt64 COMMENT 'Number of replicas which was chosen, but found to be unavailable during query execution with task-based parallel replicas', `ProfileEvent_SharedMergeTreeVirtualPartsUpdates` UInt64 COMMENT 'Virtual parts update count', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesByLeader` UInt64 COMMENT 'Virtual parts updates by leader', `ProfileEvent_SharedMergeTreeVirtualPartsUpdateMicroseconds` UInt64 COMMENT 'Virtual parts update microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromZooKeeper` UInt64 COMMENT 'Virtual parts updates count from ZooKeeper', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromZooKeeperMicroseconds` UInt64 COMMENT 'Virtual parts updates from ZooKeeper microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesPeerNotFound` UInt64 COMMENT 'Virtual updates from peer failed because no one found', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromPeer` UInt64 COMMENT 'Virtual parts updates count from peer', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromPeerMicroseconds` UInt64 COMMENT 'Virtual parts updates from peer microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesForMergesOrStatus` UInt64 COMMENT 'Virtual parts updates from non-default background job', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesLeaderFailedElection` UInt64 COMMENT 'Virtual parts updates leader election failed', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesLeaderSuccessfulElection` UInt64 COMMENT 'Virtual parts updates leader election successful', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentAttempt` UInt64 COMMENT 'How many times we tried to assign merge or mutation', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentFailedWithNothingToDo` UInt64 COMMENT 'How many times we tried to assign merge or mutation and failed because nothing to merge', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentFailedWithConflict` UInt64 COMMENT 'How many times we tried to assign merge or mutation and failed because of conflict in Keeper', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentSuccessful` UInt64 COMMENT 'How many times we tried to assign merge or mutation', `ProfileEvent_SharedMergeTreeMergePartsMovedToOudated` UInt64 COMMENT 'How many parts moved to oudated directory', `ProfileEvent_SharedMergeTreeMergePartsMovedToCondemned` UInt64 COMMENT 'How many parts moved to condemned directory', `ProfileEvent_SharedMergeTreeOutdatedPartsConfirmationRequest` UInt64 COMMENT 'How many ZooKeeper requests were used to config outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsConfirmationInvocations` UInt64 COMMENT 'How many invocations were made to confirm outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsHTTPRequest` UInt64 COMMENT 'How many HTTP requests were send to confirm outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsHTTPResponse` UInt64 COMMENT 'How many HTTP responses were send to confirm outdated parts', `ProfileEvent_SharedMergeTreeCondemnedPartsKillRequest` UInt64 COMMENT 'How many ZooKeeper requests were used to remove condemned parts', `ProfileEvent_SharedMergeTreeCondemnedPartsLockConfict` UInt64 COMMENT 'How many times we failed to acquite lock because of conflict', `ProfileEvent_SharedMergeTreeCondemnedPartsRemoved` UInt64 COMMENT 'How many condemned parts were removed', `ProfileEvent_KeeperLogsEntryReadFromLatestCache` UInt64 COMMENT 'Number of log entries in Keeper being read from latest logs cache', `ProfileEvent_KeeperLogsEntryReadFromCommitCache` UInt64 COMMENT 'Number of log entries in Keeper being read from commit logs cache', `ProfileEvent_KeeperLogsEntryReadFromFile` UInt64 COMMENT 'Number of log entries in Keeper being read directly from the changelog file', `ProfileEvent_KeeperLogsPrefetchedEntries` UInt64 COMMENT 'Number of log entries in Keeper being prefetched from the changelog file', `ProfileEvent_StorageConnectionsCreated` UInt64 COMMENT 'Number of created connections for storages', `ProfileEvent_StorageConnectionsReused` UInt64 COMMENT 'Number of reused connections for storages', `ProfileEvent_StorageConnectionsReset` UInt64 COMMENT 'Number of reset connections for storages', `ProfileEvent_StorageConnectionsPreserved` UInt64 COMMENT 'Number of preserved connections for storages', `ProfileEvent_StorageConnectionsExpired` UInt64 COMMENT 'Number of expired connections for storages', `ProfileEvent_StorageConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a connection for storage is failed', `ProfileEvent_StorageConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating connections for storages', `ProfileEvent_DiskConnectionsCreated` UInt64 COMMENT 'Number of created connections for disk', `ProfileEvent_DiskConnectionsReused` UInt64 COMMENT 'Number of reused connections for disk', `ProfileEvent_DiskConnectionsReset` UInt64 COMMENT 'Number of reset connections for disk', `ProfileEvent_DiskConnectionsPreserved` UInt64 COMMENT 'Number of preserved connections for disk', `ProfileEvent_DiskConnectionsExpired` UInt64 COMMENT 'Number of expired connections for disk', `ProfileEvent_DiskConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a connection for disk is failed', `ProfileEvent_DiskConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating connections for disk', `ProfileEvent_HTTPConnectionsCreated` UInt64 COMMENT 'Number of created http connections', `ProfileEvent_HTTPConnectionsReused` UInt64 COMMENT 'Number of reused http connections', `ProfileEvent_HTTPConnectionsReset` UInt64 COMMENT 'Number of reset http connections', `ProfileEvent_HTTPConnectionsPreserved` UInt64 COMMENT 'Number of preserved http connections', `ProfileEvent_HTTPConnectionsExpired` UInt64 COMMENT 'Number of expired http connections', `ProfileEvent_HTTPConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a http connection failed', `ProfileEvent_HTTPConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating http connections', `ProfileEvent_AddressesDiscovered` UInt64 COMMENT 'Total count of new addresses in dns resolve results for http connections', `ProfileEvent_AddressesExpired` UInt64 COMMENT 'Total count of expired addresses which is no longer presented in dns resolve results for http connections', `ProfileEvent_AddressesMarkedAsFailed` UInt64 COMMENT 'Total count of addresses which has been marked as faulty due to connection errors for http connections', `ProfileEvent_ReadWriteBufferFromHTTPRequestsSent` UInt64 COMMENT 'Number of HTTP requests sent by ReadWriteBufferFromHTTP', `ProfileEvent_ReadWriteBufferFromHTTPBytes` UInt64 COMMENT 'Total size of payload bytes received and sent by ReadWriteBufferFromHTTP. Doesn\'t include HTTP headers.', `ProfileEvent_ConcurrencyControlSlotsGranted` UInt64 COMMENT 'Number of CPU slot granted according to guarantee of 1 thread per query and for queries with setting \'use_concurrency_control\' = 0', `ProfileEvent_ConcurrencyControlSlotsDelayed` UInt64 COMMENT 'Number of CPU slot not granted initially and required to wait for a free CPU slot', `ProfileEvent_ConcurrencyControlSlotsAcquired` UInt64 COMMENT 'Total number of CPU slot acquired', `ProfileEvent_ConcurrencyControlSlotsAcquiredNonCompeting` UInt64 COMMENT 'Total number of noncompeting CPU slot acquired', `ProfileEvent_ConcurrencyControlQueriesDelayed` UInt64 COMMENT 'Total number of CPU slot allocations (queries) that were required to wait for slots to upscale', `ProfileEvent_SharedDatabaseCatalogFailedToApplyState` UInt64 COMMENT 'Number of failures to apply new state in SharedDatabaseCatalog', `ProfileEvent_SharedDatabaseCatalogStateApplicationMicroseconds` UInt64 COMMENT 'Total time spend on application of new state in SharedDatabaseCatalog', `ProfileEvent_GWPAsanAllocateSuccess` UInt64 COMMENT 'Number of successful allocations done by GWPAsan', `ProfileEvent_GWPAsanAllocateFailed` UInt64 COMMENT 'Number of failed allocations done by GWPAsan (i.e. filled pool)', `ProfileEvent_GWPAsanFree` UInt64 COMMENT 'Number of free operations done by GWPAsan', `ProfileEvent_MemoryWorkerRun` UInt64 COMMENT 'Number of runs done by MemoryWorker in background', `ProfileEvent_MemoryWorkerRunElapsedMicroseconds` UInt64 COMMENT 'Total time spent by MemoryWorker for background work', `ProfileEvent_ParquetFetchWaitTimeMicroseconds` UInt64 COMMENT 'Time of waiting fetching parquet data', `CurrentMetric_Query` Int64 COMMENT 'Number of executing queries', `CurrentMetric_Merge` Int64 COMMENT 'Number of executing background merges', `CurrentMetric_MergeParts` Int64 COMMENT 'Number of source parts participating in current background merges', `CurrentMetric_Move` Int64 COMMENT 'Number of currently executing moves', `CurrentMetric_PartMutation` Int64 COMMENT 'Number of mutations (ALTER DELETE/UPDATE)', `CurrentMetric_ReplicatedFetch` Int64 COMMENT 'Number of data parts being fetched from replica', `CurrentMetric_ReplicatedSend` Int64 COMMENT 'Number of data parts being sent to replicas', `CurrentMetric_ReplicatedChecks` Int64 COMMENT 'Number of data parts checking for consistency', `CurrentMetric_BackgroundMergesAndMutationsPoolTask` Int64 COMMENT 'Number of active merges and mutations in an associated background pool', `CurrentMetric_BackgroundMergesAndMutationsPoolSize` Int64 COMMENT 'Limit on number of active merges and mutations in an associated background pool', `CurrentMetric_BackgroundFetchesPoolTask` Int64 COMMENT 'Number of active fetches in an associated background pool', `CurrentMetric_BackgroundFetchesPoolSize` Int64 COMMENT 'Limit on number of simultaneous fetches in an associated background pool', `CurrentMetric_BackgroundCommonPoolTask` Int64 COMMENT 'Number of active tasks in an associated background pool', `CurrentMetric_BackgroundCommonPoolSize` Int64 COMMENT 'Limit on number of tasks in an associated background pool', `CurrentMetric_BackgroundMovePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundProcessingPool for moves', `CurrentMetric_BackgroundMovePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundProcessingPool for moves', `CurrentMetric_BackgroundSchedulePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundSchedulePool. This pool is used for periodic ReplicatedMergeTree tasks, like cleaning old data parts, altering data parts, replica re-initialization, etc.', `CurrentMetric_BackgroundSchedulePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundSchedulePool. This pool is used for periodic ReplicatedMergeTree tasks, like cleaning old data parts, altering data parts, replica re-initialization, etc.', `CurrentMetric_BackgroundBufferFlushSchedulePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundBufferFlushSchedulePool. This pool is used for periodic Buffer flushes', `CurrentMetric_BackgroundBufferFlushSchedulePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundBufferFlushSchedulePool', `CurrentMetric_BackgroundDistributedSchedulePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundDistributedSchedulePool. This pool is used for distributed sends that is done in background.', `CurrentMetric_BackgroundDistributedSchedulePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundDistributedSchedulePool', `CurrentMetric_BackgroundMessageBrokerSchedulePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundMessageBrokerSchedulePool for message streaming', `CurrentMetric_BackgroundMessageBrokerSchedulePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundMessageBrokerSchedulePool for message streaming', `CurrentMetric_CacheDictionaryUpdateQueueBatches` Int64 COMMENT 'Number of \'batches\' (a set of keys) in update queue in CacheDictionaries.', `CurrentMetric_CacheDictionaryUpdateQueueKeys` Int64 COMMENT 'Exact number of keys in update queue in CacheDictionaries.', `CurrentMetric_DiskSpaceReservedForMerge` Int64 COMMENT 'Disk space reserved for currently running background merges. It is slightly more than the total size of currently merging parts.', `CurrentMetric_DistributedSend` Int64 COMMENT 'Number of connections to remote servers sending data that was INSERTed into Distributed tables. Both synchronous and asynchronous mode.', `CurrentMetric_QueryPreempted` Int64 COMMENT 'Number of queries that are stopped and waiting due to \'priority\' setting.', `CurrentMetric_TCPConnection` Int64 COMMENT 'Number of connections to TCP server (clients with native interface), also included server-server distributed query connections', `CurrentMetric_MySQLConnection` Int64 COMMENT 'Number of client connections using MySQL protocol', `CurrentMetric_HTTPConnection` Int64 COMMENT 'Number of connections to HTTP server', `CurrentMetric_InterserverConnection` Int64 COMMENT 'Number of connections from other replicas to fetch parts', `CurrentMetric_PostgreSQLConnection` Int64 COMMENT 'Number of client connections using PostgreSQL protocol', `CurrentMetric_OpenFileForRead` Int64 COMMENT 'Number of files open for reading', `CurrentMetric_OpenFileForWrite` Int64 COMMENT 'Number of files open for writing', `CurrentMetric_Compressing` Int64 COMMENT 'Number of compress operations using internal compression codecs', `CurrentMetric_Decompressing` Int64 COMMENT 'Number of decompress operations using internal compression codecs', `CurrentMetric_ParallelCompressedWriteBufferThreads` Int64 COMMENT 'Number of threads in all instances of ParallelCompressedWriteBuffer - these threads are doing parallel compression and writing', `CurrentMetric_ParallelCompressedWriteBufferWait` Int64 COMMENT 'Number of threads in all instances of ParallelCompressedWriteBuffer that are currently waiting for buffer to become available for writing', `CurrentMetric_TotalTemporaryFiles` Int64 COMMENT 'Number of temporary files created', `CurrentMetric_TemporaryFilesForSort` Int64 COMMENT 'Number of temporary files created for external sorting', `CurrentMetric_TemporaryFilesForAggregation` Int64 COMMENT 'Number of temporary files created for external aggregation', `CurrentMetric_TemporaryFilesForJoin` Int64 COMMENT 'Number of temporary files created for JOIN', `CurrentMetric_TemporaryFilesForMerge` Int64 COMMENT 'Number of temporary files for vertical merge', `CurrentMetric_TemporaryFilesUnknown` Int64 COMMENT 'Number of temporary files created without known purpose', `CurrentMetric_Read` Int64 COMMENT 'Number of read (read, pread, io_getevents, etc.) syscalls in fly', `CurrentMetric_RemoteRead` Int64 COMMENT 'Number of read with remote reader in fly', `CurrentMetric_Write` Int64 COMMENT 'Number of write (write, pwrite, io_getevents, etc.) syscalls in fly', `CurrentMetric_NetworkReceive` Int64 COMMENT 'Number of threads receiving data from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `CurrentMetric_NetworkSend` Int64 COMMENT 'Number of threads sending data to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `CurrentMetric_SendScalars` Int64 COMMENT 'Number of connections that are sending data for scalars to remote servers.', `CurrentMetric_SendExternalTables` Int64 COMMENT 'Number of connections that are sending data for external tables to remote servers. External tables are used to implement GLOBAL IN and GLOBAL JOIN operators with distributed subqueries.', `CurrentMetric_QueryThread` Int64 COMMENT 'Number of query processing threads', `CurrentMetric_ReadonlyReplica` Int64 COMMENT 'Number of Replicated tables that are currently in readonly state due to re-initialization after ZooKeeper session loss or due to startup without ZooKeeper configured.', `CurrentMetric_MemoryTracking` Int64 COMMENT 'Total amount of memory (bytes) allocated by the server.', `CurrentMetric_MemoryTrackingUncorrected` Int64 COMMENT 'Total amount of memory (bytes) allocated by the server not corrected by RSS.', `CurrentMetric_MergesMutationsMemoryTracking` Int64 COMMENT 'Total amount of memory (bytes) allocated by background tasks (merges and mutations).', `CurrentMetric_EphemeralNode` Int64 COMMENT 'Number of ephemeral nodes hold in ZooKeeper.', `CurrentMetric_ZooKeeperSession` Int64 COMMENT 'Number of sessions (connections) to ZooKeeper. Should be no more than one, because using more than one connection to ZooKeeper may lead to bugs due to lack of linearizability (stale reads) that ZooKeeper consistency model allows.', `CurrentMetric_ZooKeeperWatch` Int64 COMMENT 'Number of watches (event subscriptions) in ZooKeeper.', `CurrentMetric_ZooKeeperRequest` Int64 COMMENT 'Number of requests to ZooKeeper in fly.', `CurrentMetric_DelayedInserts` Int64 COMMENT 'Number of INSERT queries that are throttled due to high number of active data parts for partition in a MergeTree table.', `CurrentMetric_ContextLockWait` Int64 COMMENT 'Number of threads waiting for lock in Context. This is global lock.', `CurrentMetric_StorageBufferRows` Int64 COMMENT 'Number of rows in buffers of Buffer tables', `CurrentMetric_StorageBufferBytes` Int64 COMMENT 'Number of bytes in buffers of Buffer tables', `CurrentMetric_DictCacheRequests` Int64 COMMENT 'Number of requests in fly to data sources of dictionaries of cache type.', `CurrentMetric_Revision` Int64 COMMENT 'Revision of the server. It is a number incremented for every release or release candidate except patch releases.', `CurrentMetric_VersionInteger` Int64 COMMENT 'Version of the server in a single integer number in base-1000. For example, version 11.22.33 is translated to 11022033.', `CurrentMetric_RWLockWaitingReaders` Int64 COMMENT 'Number of threads waiting for read on a table RWLock.', `CurrentMetric_RWLockWaitingWriters` Int64 COMMENT 'Number of threads waiting for write on a table RWLock.', `CurrentMetric_RWLockActiveReaders` Int64 COMMENT 'Number of threads holding read lock in a table RWLock.', `CurrentMetric_RWLockActiveWriters` Int64 COMMENT 'Number of threads holding write lock in a table RWLock.', `CurrentMetric_GlobalThread` Int64 COMMENT 'Number of threads in global thread pool.', `CurrentMetric_GlobalThreadActive` Int64 COMMENT 'Number of threads in global thread pool running a task.', `CurrentMetric_GlobalThreadScheduled` Int64 COMMENT 'Number of queued or active jobs in global thread pool.', `CurrentMetric_LocalThread` Int64 COMMENT 'Obsolete. Number of threads in local thread pools. The threads in local thread pools are taken from the global thread pool.', `CurrentMetric_LocalThreadActive` Int64 COMMENT 'Obsolete. Number of threads in local thread pools running a task.', `CurrentMetric_LocalThreadScheduled` Int64 COMMENT 'Obsolete. Number of queued or active jobs in local thread pools.', `CurrentMetric_MergeTreeDataSelectExecutorThreads` Int64 COMMENT 'Number of threads in the MergeTreeDataSelectExecutor thread pool.', `CurrentMetric_MergeTreeDataSelectExecutorThreadsActive` Int64 COMMENT 'Number of threads in the MergeTreeDataSelectExecutor thread pool running a task.', `CurrentMetric_MergeTreeDataSelectExecutorThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the MergeTreeDataSelectExecutor thread pool.', `CurrentMetric_BackupsThreads` Int64 COMMENT 'Number of threads in the thread pool for BACKUP.', `CurrentMetric_BackupsThreadsActive` Int64 COMMENT 'Number of threads in thread pool for BACKUP running a task.', `CurrentMetric_BackupsThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs for BACKUP.', `CurrentMetric_RestoreThreads` Int64 COMMENT 'Number of threads in the thread pool for RESTORE.', `CurrentMetric_RestoreThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for RESTORE running a task.', `CurrentMetric_RestoreThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs for RESTORE.', `CurrentMetric_MarksLoaderThreads` Int64 COMMENT 'Number of threads in thread pool for loading marks.', `CurrentMetric_MarksLoaderThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for loading marks running a task.', `CurrentMetric_MarksLoaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for loading marks.', `CurrentMetric_IOPrefetchThreads` Int64 COMMENT 'Number of threads in the IO prefetch thread pool.', `CurrentMetric_IOPrefetchThreadsActive` Int64 COMMENT 'Number of threads in the IO prefetch thread pool running a task.', `CurrentMetric_IOPrefetchThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the IO prefetch thread pool.', `CurrentMetric_IOWriterThreads` Int64 COMMENT 'Number of threads in the IO writer thread pool.', `CurrentMetric_IOWriterThreadsActive` Int64 COMMENT 'Number of threads in the IO writer thread pool running a task.', `CurrentMetric_IOWriterThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the IO writer thread pool.', `CurrentMetric_IOThreads` Int64 COMMENT 'Number of threads in the IO thread pool.', `CurrentMetric_IOThreadsActive` Int64 COMMENT 'Number of threads in the IO thread pool running a task.', `CurrentMetric_IOThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the IO thread pool.', `CurrentMetric_CompressionThread` Int64 COMMENT 'Number of threads in compression thread pools.', `CurrentMetric_CompressionThreadActive` Int64 COMMENT 'Number of threads in compression thread pools running a task.', `CurrentMetric_CompressionThreadScheduled` Int64 COMMENT 'Number of queued or active jobs in compression thread pools.', `CurrentMetric_ThreadPoolRemoteFSReaderThreads` Int64 COMMENT 'Number of threads in the thread pool for remote_filesystem_read_method=threadpool.', `CurrentMetric_ThreadPoolRemoteFSReaderThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for remote_filesystem_read_method=threadpool running a task.', `CurrentMetric_ThreadPoolRemoteFSReaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for remote_filesystem_read_method=threadpool.', `CurrentMetric_ThreadPoolFSReaderThreads` Int64 COMMENT 'Number of threads in the thread pool for local_filesystem_read_method=threadpool.', `CurrentMetric_ThreadPoolFSReaderThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for local_filesystem_read_method=threadpool running a task.', `CurrentMetric_ThreadPoolFSReaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for local_filesystem_read_method=threadpool.', `CurrentMetric_BackupsIOThreads` Int64 COMMENT 'Number of threads in the BackupsIO thread pool.', `CurrentMetric_BackupsIOThreadsActive` Int64 COMMENT 'Number of threads in the BackupsIO thread pool running a task.', `CurrentMetric_BackupsIOThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the BackupsIO thread pool.', `CurrentMetric_DiskObjectStorageAsyncThreads` Int64 COMMENT 'Obsolete metric, shows nothing.', `CurrentMetric_DiskObjectStorageAsyncThreadsActive` Int64 COMMENT 'Obsolete metric, shows nothing.', `CurrentMetric_StorageHiveThreads` Int64 COMMENT 'Number of threads in the StorageHive thread pool.', `CurrentMetric_StorageHiveThreadsActive` Int64 COMMENT 'Number of threads in the StorageHive thread pool running a task.', `CurrentMetric_StorageHiveThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the StorageHive thread pool.', `CurrentMetric_TablesLoaderBackgroundThreads` Int64 COMMENT 'Number of threads in the tables loader background thread pool.', `CurrentMetric_TablesLoaderBackgroundThreadsActive` Int64 COMMENT 'Number of threads in the tables loader background thread pool running a task.', `CurrentMetric_TablesLoaderBackgroundThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the tables loader background thread pool.', `CurrentMetric_TablesLoaderForegroundThreads` Int64 COMMENT 'Number of threads in the tables loader foreground thread pool.', `CurrentMetric_TablesLoaderForegroundThreadsActive` Int64 COMMENT 'Number of threads in the tables loader foreground thread pool running a task.', `CurrentMetric_TablesLoaderForegroundThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the tables loader foreground thread pool.', `CurrentMetric_DatabaseOnDiskThreads` Int64 COMMENT 'Number of threads in the DatabaseOnDisk thread pool.', `CurrentMetric_DatabaseOnDiskThreadsActive` Int64 COMMENT 'Number of threads in the DatabaseOnDisk thread pool running a task.', `CurrentMetric_DatabaseOnDiskThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DatabaseOnDisk thread pool.', `CurrentMetric_DatabaseBackupThreads` Int64 COMMENT 'Number of threads in the DatabaseBackup thread pool.', `CurrentMetric_DatabaseBackupThreadsActive` Int64 COMMENT 'Number of threads in the DatabaseBackup thread pool running a task.', `CurrentMetric_DatabaseBackupThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DatabaseBackup thread pool.', `CurrentMetric_DatabaseCatalogThreads` Int64 COMMENT 'Number of threads in the DatabaseCatalog thread pool.', `CurrentMetric_DatabaseCatalogThreadsActive` Int64 COMMENT 'Number of threads in the DatabaseCatalog thread pool running a task.', `CurrentMetric_DatabaseCatalogThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DatabaseCatalog thread pool.', `CurrentMetric_DestroyAggregatesThreads` Int64 COMMENT 'Number of threads in the thread pool for destroy aggregate states.', `CurrentMetric_DestroyAggregatesThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for destroy aggregate states running a task.', `CurrentMetric_DestroyAggregatesThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for destroy aggregate states.', `CurrentMetric_ConcurrentHashJoinPoolThreads` Int64 COMMENT 'Number of threads in the thread pool for concurrent hash join.', `CurrentMetric_ConcurrentHashJoinPoolThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for concurrent hash join running a task.', `CurrentMetric_ConcurrentHashJoinPoolThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for concurrent hash join.', `CurrentMetric_HashedDictionaryThreads` Int64 COMMENT 'Number of threads in the HashedDictionary thread pool.', `CurrentMetric_HashedDictionaryThreadsActive` Int64 COMMENT 'Number of threads in the HashedDictionary thread pool running a task.', `CurrentMetric_HashedDictionaryThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the HashedDictionary thread pool.', `CurrentMetric_CacheDictionaryThreads` Int64 COMMENT 'Number of threads in the CacheDictionary thread pool.', `CurrentMetric_CacheDictionaryThreadsActive` Int64 COMMENT 'Number of threads in the CacheDictionary thread pool running a task.', `CurrentMetric_CacheDictionaryThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the CacheDictionary thread pool.', `CurrentMetric_ParallelFormattingOutputFormatThreads` Int64 COMMENT 'Number of threads in the ParallelFormattingOutputFormatThreads thread pool.', `CurrentMetric_ParallelFormattingOutputFormatThreadsActive` Int64 COMMENT 'Number of threads in the ParallelFormattingOutputFormatThreads thread pool running a task.', `CurrentMetric_ParallelFormattingOutputFormatThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the ParallelFormattingOutputFormatThreads thread pool.', `CurrentMetric_ParallelParsingInputFormatThreads` Int64 COMMENT 'Number of threads in the ParallelParsingInputFormat thread pool.', `CurrentMetric_ParallelParsingInputFormatThreadsActive` Int64 COMMENT 'Number of threads in the ParallelParsingInputFormat thread pool running a task.', `CurrentMetric_ParallelParsingInputFormatThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the ParallelParsingInputFormat thread pool.', `CurrentMetric_MergeTreeBackgroundExecutorThreads` Int64 COMMENT 'Number of threads in the MergeTreeBackgroundExecutor thread pool.', `CurrentMetric_MergeTreeBackgroundExecutorThreadsActive` Int64 COMMENT 'Number of threads in the MergeTreeBackgroundExecutor thread pool running a task.', `CurrentMetric_MergeTreeBackgroundExecutorThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the MergeTreeBackgroundExecutor thread pool.', `CurrentMetric_AsynchronousInsertThreads` Int64 COMMENT 'Number of threads in the AsynchronousInsert thread pool.', `CurrentMetric_AsynchronousInsertThreadsActive` Int64 COMMENT 'Number of threads in the AsynchronousInsert thread pool running a task.', `CurrentMetric_AsynchronousInsertThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the AsynchronousInsert thread pool.', `CurrentMetric_AsynchronousInsertQueueSize` Int64 COMMENT 'Number of pending tasks in the AsynchronousInsert queue.', `CurrentMetric_AsynchronousInsertQueueBytes` Int64 COMMENT 'Number of pending bytes in the AsynchronousInsert queue.', `CurrentMetric_StartupSystemTablesThreads` Int64 COMMENT 'Number of threads in the StartupSystemTables thread pool.', `CurrentMetric_StartupSystemTablesThreadsActive` Int64 COMMENT 'Number of threads in the StartupSystemTables thread pool running a task.', `CurrentMetric_StartupSystemTablesThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the StartupSystemTables thread pool.', `CurrentMetric_AggregatorThreads` Int64 COMMENT 'Number of threads in the Aggregator thread pool.', `CurrentMetric_AggregatorThreadsActive` Int64 COMMENT 'Number of threads in the Aggregator thread pool running a task.', `CurrentMetric_AggregatorThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the Aggregator thread pool.', `CurrentMetric_DDLWorkerThreads` Int64 COMMENT 'Number of threads in the DDLWorker thread pool for ON CLUSTER queries.', `CurrentMetric_DDLWorkerThreadsActive` Int64 COMMENT 'Number of threads in the DDLWORKER thread pool for ON CLUSTER queries running a task.', `CurrentMetric_DDLWorkerThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DDLWORKER thread pool for ON CLUSTER queries.', `CurrentMetric_StorageDistributedThreads` Int64 COMMENT 'Number of threads in the StorageDistributed thread pool.', `CurrentMetric_StorageDistributedThreadsActive` Int64 COMMENT 'Number of threads in the StorageDistributed thread pool running a task.', `CurrentMetric_StorageDistributedThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the StorageDistributed thread pool.', `CurrentMetric_DistributedInsertThreads` Int64 COMMENT 'Number of threads used for INSERT into Distributed.', `CurrentMetric_DistributedInsertThreadsActive` Int64 COMMENT 'Number of threads used for INSERT into Distributed running a task.', `CurrentMetric_DistributedInsertThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs used for INSERT into Distributed.', `CurrentMetric_StorageS3Threads` Int64 COMMENT 'Number of threads in the StorageS3 thread pool.', `CurrentMetric_StorageS3ThreadsActive` Int64 COMMENT 'Number of threads in the StorageS3 thread pool running a task.', `CurrentMetric_StorageS3ThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the StorageS3 thread pool.', `CurrentMetric_ObjectStorageS3Threads` Int64 COMMENT 'Number of threads in the S3ObjectStorage thread pool.', `CurrentMetric_ObjectStorageS3ThreadsActive` Int64 COMMENT 'Number of threads in the S3ObjectStorage thread pool running a task.', `CurrentMetric_ObjectStorageS3ThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the S3ObjectStorage thread pool.', `CurrentMetric_StorageObjectStorageThreads` Int64 COMMENT 'Number of threads in the remote table engines thread pools.', `CurrentMetric_StorageObjectStorageThreadsActive` Int64 COMMENT 'Number of threads in the remote table engines thread pool running a task.', `CurrentMetric_StorageObjectStorageThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in remote table engines thread pool.', `CurrentMetric_ObjectStorageAzureThreads` Int64 COMMENT 'Number of threads in the AzureObjectStorage thread pool.', `CurrentMetric_ObjectStorageAzureThreadsActive` Int64 COMMENT 'Number of threads in the AzureObjectStorage thread pool running a task.', `CurrentMetric_ObjectStorageAzureThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the AzureObjectStorage thread pool.', `CurrentMetric_BuildVectorSimilarityIndexThreads` Int64 COMMENT 'Number of threads in the build vector similarity index thread pool.', `CurrentMetric_BuildVectorSimilarityIndexThreadsActive` Int64 COMMENT 'Number of threads in the build vector similarity index thread pool running a task.', `CurrentMetric_BuildVectorSimilarityIndexThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the build vector similarity index thread pool.', `CurrentMetric_ObjectStorageQueueRegisteredServers` Int64 COMMENT 'Number of registered servers in StorageS3(Azure)Queue', `CurrentMetric_IcebergCatalogThreads` Int64 COMMENT 'Number of threads in the IcebergCatalog thread pool.', `CurrentMetric_IcebergCatalogThreadsActive` Int64 COMMENT 'Number of threads in the IcebergCatalog thread pool running a task.', `CurrentMetric_IcebergCatalogThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the IcebergCatalog thread pool.', `CurrentMetric_ParallelWithQueryThreads` Int64 COMMENT 'Number of threads in the threadpool for processing PARALLEL WITH queries.', `CurrentMetric_ParallelWithQueryActiveThreads` Int64 COMMENT 'Number of active threads in the threadpool for processing PARALLEL WITH queries.', `CurrentMetric_ParallelWithQueryScheduledThreads` Int64 COMMENT 'Number of queued or active jobs in the threadpool for processing PARALLEL WITH queries.', `CurrentMetric_DiskPlainRewritableAzureDirectoryMapSize` Int64 COMMENT 'Number of local-to-remote path entries in the \'plain_rewritable\' in-memory map for AzureObjectStorage.', `CurrentMetric_DiskPlainRewritableAzureFileCount` Int64 COMMENT 'Number of file entries in the \'plain_rewritable\' in-memory map for AzureObjectStorage.', `CurrentMetric_DiskPlainRewritableAzureUniqueFileNamesCount` Int64 COMMENT 'Number of unique file name entries in the \'plain_rewritable\' in-memory map for AzureObjectStorage.', `CurrentMetric_DiskPlainRewritableLocalDirectoryMapSize` Int64 COMMENT 'Number of local-to-remote path entries in the \'plain_rewritable\' in-memory map for LocalObjectStorage.', `CurrentMetric_DiskPlainRewritableLocalFileCount` Int64 COMMENT 'Number of file entries in the \'plain_rewritable\' in-memory map for LocalObjectStorage.', `CurrentMetric_DiskPlainRewritableLocalUniqueFileNamesCount` Int64 COMMENT 'Number of unique file name entries in the \'plain_rewritable\' in-memory map for LocalObjectStorage.', `CurrentMetric_DiskPlainRewritableS3DirectoryMapSize` Int64 COMMENT 'Number of local-to-remote path entries in the \'plain_rewritable\' in-memory map for S3ObjectStorage.', `CurrentMetric_DiskPlainRewritableS3FileCount` Int64 COMMENT 'Number of file entries in the \'plain_rewritable\' in-memory map for S3ObjectStorage.', `CurrentMetric_DiskPlainRewritableS3UniqueFileNamesCount` Int64 COMMENT 'Number of unique file name entries in the \'plain_rewritable\' in-memory map for S3ObjectStorage.', `CurrentMetric_MergeTreeFetchPartitionThreads` Int64 COMMENT 'Number of threads for ALTER TABLE FETCH PARTITION', `CurrentMetric_MergeTreeFetchPartitionThreadsActive` Int64 COMMENT 'Number of threads for ALTER TABLE FETCH PARTITION fetching part', `CurrentMetric_MergeTreeFetchPartitionThreadsScheduled` Int64 COMMENT 'Number of queued or active part fetches in ALTER TABLE FETCH PARTITION', `CurrentMetric_MergeTreePartsLoaderThreads` Int64 COMMENT 'Number of threads in the MergeTree parts loader thread pool.', `CurrentMetric_MergeTreePartsLoaderThreadsActive` Int64 COMMENT 'Number of threads in the MergeTree parts loader thread pool running a task.', `CurrentMetric_MergeTreePartsLoaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the MergeTree parts loader thread pool.', `CurrentMetric_MergeTreeOutdatedPartsLoaderThreads` Int64 COMMENT 'Number of threads in the threadpool for loading Outdated data parts.', `CurrentMetric_MergeTreeOutdatedPartsLoaderThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for loading Outdated data parts.', `CurrentMetric_MergeTreeOutdatedPartsLoaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for loading Outdated data parts.', `CurrentMetric_MergeTreeUnexpectedPartsLoaderThreads` Int64 COMMENT 'Number of threads in the threadpool for loading Unexpected data parts.', `CurrentMetric_MergeTreeUnexpectedPartsLoaderThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for loading Unexpected data parts.', `CurrentMetric_MergeTreeUnexpectedPartsLoaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for loading Unexpected data parts.', `CurrentMetric_MergeTreePartsCleanerThreads` Int64 COMMENT 'Number of threads in the MergeTree parts cleaner thread pool.', `CurrentMetric_MergeTreePartsCleanerThreadsActive` Int64 COMMENT 'Number of threads in the MergeTree parts cleaner thread pool running a task.', `CurrentMetric_MergeTreePartsCleanerThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the MergeTree parts cleaner thread pool.', `CurrentMetric_DatabaseReplicatedCreateTablesThreads` Int64 COMMENT 'Number of threads in the threadpool for table creation in DatabaseReplicated.', `CurrentMetric_DatabaseReplicatedCreateTablesThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for table creation in DatabaseReplicated.', `CurrentMetric_DatabaseReplicatedCreateTablesThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for table creation in DatabaseReplicated.', `CurrentMetric_IDiskCopierThreads` Int64 COMMENT 'Number of threads for copying data between disks of different types.', `CurrentMetric_IDiskCopierThreadsActive` Int64 COMMENT 'Number of threads for copying data between disks of different types running a task.', `CurrentMetric_IDiskCopierThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs for copying data between disks of different types.', `CurrentMetric_SystemReplicasThreads` Int64 COMMENT 'Number of threads in the system.replicas thread pool.', `CurrentMetric_SystemReplicasThreadsActive` Int64 COMMENT 'Number of threads in the system.replicas thread pool running a task.', `CurrentMetric_SystemReplicasThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the system.replicas thread pool.', `CurrentMetric_RestartReplicaThreads` Int64 COMMENT 'Number of threads in the RESTART REPLICA thread pool.', `CurrentMetric_RestartReplicaThreadsActive` Int64 COMMENT 'Number of threads in the RESTART REPLICA thread pool running a task.', `CurrentMetric_RestartReplicaThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the RESTART REPLICA thread pool.', `CurrentMetric_QueryPipelineExecutorThreads` Int64 COMMENT 'Number of threads in the PipelineExecutor thread pool.', `CurrentMetric_QueryPipelineExecutorThreadsActive` Int64 COMMENT 'Number of threads in the PipelineExecutor thread pool running a task.', `CurrentMetric_QueryPipelineExecutorThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the PipelineExecutor thread pool.', `CurrentMetric_ParquetDecoderThreads` Int64 COMMENT 'Number of threads in the ParquetBlockInputFormat thread pool.', `CurrentMetric_ParquetDecoderThreadsActive` Int64 COMMENT 'Number of threads in the ParquetBlockInputFormat thread pool running a task.', `CurrentMetric_ParquetDecoderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the ParquetBlockInputFormat thread pool.', `CurrentMetric_ParquetDecoderIOThreads` Int64 COMMENT 'Number of threads in the ParquetBlockInputFormat io thread pool.', `CurrentMetric_ParquetDecoderIOThreadsActive` Int64 COMMENT 'Number of threads in the ParquetBlockInputFormat io thread pool running a task.', `CurrentMetric_ParquetDecoderIOThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the ParquetBlockInputFormat io thread pool.', `CurrentMetric_ParquetEncoderThreads` Int64 COMMENT 'Number of threads in ParquetBlockOutputFormat thread pool.', `CurrentMetric_ParquetEncoderThreadsActive` Int64 COMMENT 'Number of threads in ParquetBlockOutputFormat thread pool running a task.', `CurrentMetric_ParquetEncoderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in ParquetBlockOutputFormat thread pool.', `CurrentMetric_MergeTreeSubcolumnsReaderThreads` Int64 COMMENT 'Number of threads in the thread pool used for subcolumns reading in MergeTree.', `CurrentMetric_MergeTreeSubcolumnsReaderThreadsActive` Int64 COMMENT 'Number of threads in the thread pool used for subcolumns reading in MergeTree running a task.', `CurrentMetric_MergeTreeSubcolumnsReaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool used for subcolumns reading in MergeTree.', `CurrentMetric_DWARFReaderThreads` Int64 COMMENT 'Number of threads in the DWARFBlockInputFormat thread pool.', `CurrentMetric_DWARFReaderThreadsActive` Int64 COMMENT 'Number of threads in the DWARFBlockInputFormat thread pool running a task.', `CurrentMetric_DWARFReaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DWARFBlockInputFormat thread pool.', `CurrentMetric_OutdatedPartsLoadingThreads` Int64 COMMENT 'Number of threads in the threadpool for loading Outdated data parts.', `CurrentMetric_OutdatedPartsLoadingThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for loading Outdated data parts.', `CurrentMetric_OutdatedPartsLoadingThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for loading Outdated data parts.', `CurrentMetric_PolygonDictionaryThreads` Int64 COMMENT 'Number of threads in the threadpool for polygon dictionaries.', `CurrentMetric_PolygonDictionaryThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for polygon dictionaries.', `CurrentMetric_PolygonDictionaryThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for polygon dictionaries.', `CurrentMetric_DistributedBytesToInsert` Int64 COMMENT 'Number of pending bytes to process for asynchronous insertion into Distributed tables. Number of bytes for every shard is summed.', `CurrentMetric_BrokenDistributedBytesToInsert` Int64 COMMENT 'Number of bytes for asynchronous insertion into Distributed tables that has been marked as broken. Number of bytes for every shard is summed.', `CurrentMetric_DistributedFilesToInsert` Int64 COMMENT 'Number of pending files to process for asynchronous insertion into Distributed tables. Number of files for every shard is summed.', `CurrentMetric_BrokenDistributedFilesToInsert` Int64 COMMENT 'Number of files for asynchronous insertion into Distributed tables that has been marked as broken. Number of files for every shard is summed.', `CurrentMetric_TablesToDropQueueSize` Int64 COMMENT 'Number of dropped tables, that are waiting for background data removal.', `CurrentMetric_MaxDDLEntryID` Int64 COMMENT 'Max processed DDL entry of DDLWorker.', `CurrentMetric_MaxPushedDDLEntryID` Int64 COMMENT 'Max DDL entry of DDLWorker that pushed to zookeeper.', `CurrentMetric_PartsTemporary` Int64 COMMENT 'The part is generating now, it is not in data_parts list.', `CurrentMetric_PartsPreCommitted` Int64 COMMENT 'Deprecated. See PartsPreActive.', `CurrentMetric_PartsCommitted` Int64 COMMENT 'Deprecated. See PartsActive.', `CurrentMetric_PartsPreActive` Int64 COMMENT 'The part is in data_parts, but not used for SELECTs.', `CurrentMetric_PartsActive` Int64 COMMENT 'Active data part, used by current and upcoming SELECTs.', `CurrentMetric_AttachedDatabase` Int64 COMMENT 'Active databases.', `CurrentMetric_AttachedTable` Int64 COMMENT 'Active tables.', `CurrentMetric_AttachedReplicatedTable` Int64 COMMENT 'Active replicated tables.', `CurrentMetric_AttachedView` Int64 COMMENT 'Active views.', `CurrentMetric_AttachedDictionary` Int64 COMMENT 'Active dictionaries.', `CurrentMetric_PartsOutdated` Int64 COMMENT 'Not active data part, but could be used by only current SELECTs, could be deleted after SELECTs finishes.', `CurrentMetric_PartsDeleting` Int64 COMMENT 'Not active data part with identity refcounter, it is deleting right now by a cleaner.', `CurrentMetric_PartsDeleteOnDestroy` Int64 COMMENT 'Part was moved to another disk and should be deleted in own destructor.', `CurrentMetric_PartsWide` Int64 COMMENT 'Wide parts.', `CurrentMetric_PartsCompact` Int64 COMMENT 'Compact parts.', `CurrentMetric_MMappedFiles` Int64 COMMENT 'Total number of mmapped files.', `CurrentMetric_MMappedFileBytes` Int64 COMMENT 'Sum size of mmapped file regions.', `CurrentMetric_AsynchronousReadWait` Int64 COMMENT 'Number of threads waiting for asynchronous read.', `CurrentMetric_PendingAsyncInsert` Int64 COMMENT 'Number of asynchronous inserts that are waiting for flush.', `CurrentMetric_KafkaConsumers` Int64 COMMENT 'Number of active Kafka consumers', `CurrentMetric_KafkaConsumersWithAssignment` Int64 COMMENT 'Number of active Kafka consumers which have some partitions assigned.', `CurrentMetric_KafkaProducers` Int64 COMMENT 'Number of active Kafka producer created', `CurrentMetric_KafkaLibrdkafkaThreads` Int64 COMMENT 'Number of active librdkafka threads', `CurrentMetric_KafkaBackgroundReads` Int64 COMMENT 'Number of background reads currently working (populating materialized views from Kafka)', `CurrentMetric_KafkaConsumersInUse` Int64 COMMENT 'Number of consumers which are currently used by direct or background reads', `CurrentMetric_KafkaWrites` Int64 COMMENT 'Number of currently running inserts to Kafka', `CurrentMetric_KafkaAssignedPartitions` Int64 COMMENT 'Number of partitions Kafka tables currently assigned to', `CurrentMetric_FilesystemCacheReadBuffers` Int64 COMMENT 'Number of active cache buffers', `CurrentMetric_CacheFileSegments` Int64 COMMENT 'Number of existing cache file segments', `CurrentMetric_CacheDetachedFileSegments` Int64 COMMENT 'Number of existing detached cache file segments', `CurrentMetric_FilesystemCacheSize` Int64 COMMENT 'Filesystem cache size in bytes', `CurrentMetric_FilesystemCacheSizeLimit` Int64 COMMENT 'Filesystem cache size limit in bytes', `CurrentMetric_FilesystemCacheElements` Int64 COMMENT 'Filesystem cache elements (file segments)', `CurrentMetric_FilesystemCacheDownloadQueueElements` Int64 COMMENT 'Filesystem cache elements in download queue', `CurrentMetric_FilesystemCacheDelayedCleanupElements` Int64 COMMENT 'Filesystem cache elements in background cleanup queue', `CurrentMetric_FilesystemCacheHoldFileSegments` Int64 COMMENT 'Filesystem cache file segment which are currently hold as unreleasable', `CurrentMetric_AsyncInsertCacheSize` Int64 COMMENT 'Number of async insert hash id in cache', `CurrentMetric_SkippingIndexCacheSize` Int64 COMMENT 'Size of the skipping index cache in bytes', `CurrentMetric_S3Requests` Int64 COMMENT 'S3 requests count', `CurrentMetric_KeeperAliveConnections` Int64 COMMENT 'Number of alive connections', `CurrentMetric_KeeperOutstandingRequests` Int64 COMMENT 'Number of outstanding requests', `CurrentMetric_ThreadsInOvercommitTracker` Int64 COMMENT 'Number of waiting threads inside of OvercommitTracker', `CurrentMetric_IOUringPendingEvents` Int64 COMMENT 'Number of io_uring SQEs waiting to be submitted', `CurrentMetric_IOUringInFlightEvents` Int64 COMMENT 'Number of io_uring SQEs in flight', `CurrentMetric_ReadTaskRequestsSent` Int64 COMMENT 'The current number of callback requests in flight from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `CurrentMetric_MergeTreeReadTaskRequestsSent` Int64 COMMENT 'The current number of callback requests in flight from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `CurrentMetric_MergeTreeAllRangesAnnouncementsSent` Int64 COMMENT 'The current number of announcement being sent in flight from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `CurrentMetric_CreatedTimersInQueryProfiler` Int64 COMMENT 'Number of Created thread local timers in QueryProfiler', `CurrentMetric_ActiveTimersInQueryProfiler` Int64 COMMENT 'Number of Active thread local timers in QueryProfiler', `CurrentMetric_RefreshableViews` Int64 COMMENT 'Number materialized views with periodic refreshing (REFRESH)', `CurrentMetric_RefreshingViews` Int64 COMMENT 'Number of materialized views currently executing a refresh', `CurrentMetric_StorageBufferFlushThreads` Int64 COMMENT 'Number of threads for background flushes in StorageBuffer', `CurrentMetric_StorageBufferFlushThreadsActive` Int64 COMMENT 'Number of threads for background flushes in StorageBuffer running a task', `CurrentMetric_StorageBufferFlushThreadsScheduled` Int64 COMMENT 'Number of queued or active threads for background flushes in StorageBuffer', `CurrentMetric_SharedMergeTreeThreads` Int64 COMMENT 'Number of threads in the thread pools in internals of SharedMergeTree', `CurrentMetric_SharedMergeTreeThreadsActive` Int64 COMMENT 'Number of threads in the thread pools in internals of SharedMergeTree running a task', `CurrentMetric_SharedMergeTreeThreadsScheduled` Int64 COMMENT 'Number of queued or active threads in the thread pools in internals of SharedMergeTree', `CurrentMetric_SharedMergeTreeFetch` Int64 COMMENT 'Number of fetches in progress', `CurrentMetric_CacheWarmerBytesInProgress` Int64 COMMENT 'Total size of remote file segments waiting to be asynchronously loaded into filesystem cache.', `CurrentMetric_DistrCacheOpenedConnections` Int64 COMMENT 'Number of open connections to Distributed Cache', `CurrentMetric_DistrCacheUsedConnections` Int64 COMMENT 'Number of currently used connections to Distributed Cache', `CurrentMetric_DistrCacheAllocatedConnections` Int64 COMMENT 'Number of currently allocated connections to Distributed Cache connection pool', `CurrentMetric_DistrCacheBorrowedConnections` Int64 COMMENT 'Number of currently borrowed connections to Distributed Cache connection pool', `CurrentMetric_DistrCacheReadRequests` Int64 COMMENT 'Number of executed Read requests to Distributed Cache', `CurrentMetric_DistrCacheWriteRequests` Int64 COMMENT 'Number of executed Write requests to Distributed Cache', `CurrentMetric_DistrCacheServerConnections` Int64 COMMENT 'Number of open connections to ClickHouse server from Distributed Cache', `CurrentMetric_DistrCacheRegisteredServers` Int64 COMMENT 'Number of distributed cache registered servers', `CurrentMetric_DistrCacheRegisteredServersCurrentAZ` Int64 COMMENT 'Number of distributed cache registered servers in current az', `CurrentMetric_DistrCacheServerS3CachedClients` Int64 COMMENT 'Number of distributed cache S3 cached clients', `CurrentMetric_SchedulerIOReadScheduled` Int64 COMMENT 'Number of IO reads are being scheduled currently', `CurrentMetric_SchedulerIOWriteScheduled` Int64 COMMENT 'Number of IO writes are being scheduled currently', `CurrentMetric_StorageConnectionsStored` Int64 COMMENT 'Total count of sessions stored in the session pool for storages', `CurrentMetric_StorageConnectionsTotal` Int64 COMMENT 'Total count of all sessions: stored in the pool and actively used right now for storages', `CurrentMetric_DiskConnectionsStored` Int64 COMMENT 'Total count of sessions stored in the session pool for disks', `CurrentMetric_DiskConnectionsTotal` Int64 COMMENT 'Total count of all sessions: stored in the pool and actively used right now for disks', `CurrentMetric_HTTPConnectionsStored` Int64 COMMENT 'Total count of sessions stored in the session pool for http hosts', `CurrentMetric_HTTPConnectionsTotal` Int64 COMMENT 'Total count of all sessions: stored in the pool and actively used right now for http hosts', `CurrentMetric_AddressesActive` Int64 COMMENT 'Total count of addresses which are used for creation connections with connection pools', `CurrentMetric_AddressesBanned` Int64 COMMENT 'Total count of addresses which are banned as faulty for creation connections with connection pools', `CurrentMetric_FilteringMarksWithPrimaryKey` Int64 COMMENT 'Number of threads currently doing filtering of mark ranges by the primary key', `CurrentMetric_FilteringMarksWithSecondaryKeys` Int64 COMMENT 'Number of threads currently doing filtering of mark ranges by secondary keys', `CurrentMetric_ConcurrencyControlAcquired` Int64 COMMENT 'Total number of acquired CPU slots', `CurrentMetric_ConcurrencyControlAcquiredNonCompeting` Int64 COMMENT 'Total number of acquired CPU slots that are not considered competing (the first thread if fair_round_robin scheduler is in use)', `CurrentMetric_ConcurrencyControlSoftLimit` Int64 COMMENT 'Value of soft limit on number of CPU slots', `CurrentMetric_DiskS3NoSuchKeyErrors` Int64 COMMENT 'The number of `NoSuchKey` errors that occur when reading data from S3 cloud storage through ClickHouse disks.', `CurrentMetric_SharedCatalogStateApplicationThreads` Int64 COMMENT 'Number of threads in the threadpool for state application in Shared Catalog.', `CurrentMetric_SharedCatalogStateApplicationThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for state application in Shared Catalog.', `CurrentMetric_SharedCatalogStateApplicationThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for state application in Shared Catalog.', `CurrentMetric_SharedCatalogDropLocalThreads` Int64 COMMENT 'Number of threads in the threadpool for drop of local tables in Shared Catalog.', `CurrentMetric_SharedCatalogDropLocalThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for drop of local tables in Shared Catalog.', `CurrentMetric_SharedCatalogDropLocalThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for drop of local tables in Shared Catalog.', `CurrentMetric_SharedCatalogDropZooKeeperThreads` Int64 COMMENT 'Number of threads in the threadpool for drop of object in ZooKeeper in Shared Catalog.', `CurrentMetric_SharedCatalogDropZooKeeperThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for drop of object in ZooKeeper in Shared Catalog.', `CurrentMetric_SharedCatalogDropZooKeeperThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for drop of object in ZooKeeper in Shared Catalog.', `CurrentMetric_SharedDatabaseCatalogTablesInLocalDropDetachQueue` Int64 COMMENT 'Number of tables in the queue for local drop or detach in Shared Catalog.', `CurrentMetric_StartupScriptsExecutionState` Int64 COMMENT 'State of startup scripts execution: 0 = not finished, 1 = success, 2 = failure.', `CurrentMetric_IsServerShuttingDown` Int64 COMMENT 'Indicates if the server is shutting down: 0 = no, 1 = yes' ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains history of metrics values from tables system.metrics and system.events, periodically flushed to disk.\n\nIt is safe to truncate or drop this table at any time.' With storage policy, SHOW CREATE TABLE system.metric_log is: CREATE TABLE system.metric_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `event_date` Date COMMENT 'Event date.', `event_time` DateTime COMMENT 'Event time.', `event_time_microseconds` DateTime64(6) COMMENT 'Event time with microseconds resolution.', `ProfileEvent_Query` UInt64 COMMENT 'Number of queries to be interpreted and potentially executed. Does not include queries that failed to parse or were rejected due to AST size limits, quota limits or limits on the number of simultaneously running queries. May include internal queries initiated by ClickHouse itself. Does not count subqueries.', `ProfileEvent_SelectQuery` UInt64 COMMENT 'Same as Query, but only for SELECT queries.', `ProfileEvent_InsertQuery` UInt64 COMMENT 'Same as Query, but only for INSERT queries.', `ProfileEvent_InitialQuery` UInt64 COMMENT 'Same as Query, but only counts initial queries (see is_initial_query).', `ProfileEvent_QueriesWithSubqueries` UInt64 COMMENT 'Count queries with all subqueries', `ProfileEvent_SelectQueriesWithSubqueries` UInt64 COMMENT 'Count SELECT queries with all subqueries', `ProfileEvent_InsertQueriesWithSubqueries` UInt64 COMMENT 'Count INSERT queries with all subqueries', `ProfileEvent_SelectQueriesWithPrimaryKeyUsage` UInt64 COMMENT 'Count SELECT queries which use the primary key to evaluate the WHERE condition', `ProfileEvent_AsyncInsertQuery` UInt64 COMMENT 'Same as InsertQuery, but only for asynchronous INSERT queries.', `ProfileEvent_AsyncInsertBytes` UInt64 COMMENT 'Data size in bytes of asynchronous INSERT queries.', `ProfileEvent_AsyncInsertRows` UInt64 COMMENT 'Number of rows inserted by asynchronous INSERT queries.', `ProfileEvent_AsyncInsertCacheHits` UInt64 COMMENT 'Number of times a duplicate hash id has been found in asynchronous INSERT hash id cache.', `ProfileEvent_FailedQuery` UInt64 COMMENT 'Number of failed queries.', `ProfileEvent_FailedSelectQuery` UInt64 COMMENT 'Same as FailedQuery, but only for SELECT queries.', `ProfileEvent_FailedInsertQuery` UInt64 COMMENT 'Same as FailedQuery, but only for INSERT queries.', `ProfileEvent_FailedAsyncInsertQuery` UInt64 COMMENT 'Number of failed ASYNC INSERT queries.', `ProfileEvent_QueryTimeMicroseconds` UInt64 COMMENT 'Total time of all queries.', `ProfileEvent_SelectQueryTimeMicroseconds` UInt64 COMMENT 'Total time of SELECT queries.', `ProfileEvent_InsertQueryTimeMicroseconds` UInt64 COMMENT 'Total time of INSERT queries.', `ProfileEvent_OtherQueryTimeMicroseconds` UInt64 COMMENT 'Total time of queries that are not SELECT or INSERT.', `ProfileEvent_FileOpen` UInt64 COMMENT 'Number of files opened.', `ProfileEvent_Seek` UInt64 COMMENT 'Number of times the \'lseek\' function was called.', `ProfileEvent_ReadBufferFromFileDescriptorRead` UInt64 COMMENT 'Number of reads (read/pread) from a file descriptor. Does not include sockets.', `ProfileEvent_ReadBufferFromFileDescriptorReadFailed` UInt64 COMMENT 'Number of times the read (read/pread) from a file descriptor have failed.', `ProfileEvent_ReadBufferFromFileDescriptorReadBytes` UInt64 COMMENT 'Number of bytes read from file descriptors. If the file is compressed, this will show the compressed data size.', `ProfileEvent_WriteBufferFromFileDescriptorWrite` UInt64 COMMENT 'Number of writes (write/pwrite) to a file descriptor. Does not include sockets.', `ProfileEvent_WriteBufferFromFileDescriptorWriteFailed` UInt64 COMMENT 'Number of times the write (write/pwrite) to a file descriptor have failed.', `ProfileEvent_WriteBufferFromFileDescriptorWriteBytes` UInt64 COMMENT 'Number of bytes written to file descriptors. If the file is compressed, this will show compressed data size.', `ProfileEvent_FileSync` UInt64 COMMENT 'Number of times the F_FULLFSYNC/fsync/fdatasync function was called for files.', `ProfileEvent_DirectorySync` UInt64 COMMENT 'Number of times the F_FULLFSYNC/fsync/fdatasync function was called for directories.', `ProfileEvent_FileSyncElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for F_FULLFSYNC/fsync/fdatasync syscall for files.', `ProfileEvent_DirectorySyncElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for F_FULLFSYNC/fsync/fdatasync syscall for directories.', `ProfileEvent_ReadCompressedBytes` UInt64 COMMENT 'Number of bytes (the number of bytes before decompression) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferBlocks` UInt64 COMMENT 'Number of compressed blocks (the blocks of data that are compressed independent of each other) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferBytes` UInt64 COMMENT 'Number of uncompressed bytes (the number of bytes after decompression) read from compressed sources (files, network).', `ProfileEvent_CompressedReadBufferChecksumDoesntMatch` UInt64 COMMENT 'Number of times the compressed block checksum did not match.', `ProfileEvent_CompressedReadBufferChecksumDoesntMatchSingleBitMismatch` UInt64 COMMENT 'Number of times a compressed block checksum mismatch was caused by a single-bit difference.', `ProfileEvent_CompressedReadBufferChecksumDoesntMatchMicroseconds` UInt64 COMMENT 'Total time spent detecting bit-flips due to compressed block checksum mismatches.', `ProfileEvent_UncompressedCacheHits` UInt64 COMMENT 'Number of times a block of data has been found in the uncompressed cache (and decompression was avoided).', `ProfileEvent_UncompressedCacheMisses` UInt64 COMMENT 'Number of times a block of data has not been found in the uncompressed cache (and required decompression).', `ProfileEvent_UncompressedCacheWeightLost` UInt64 COMMENT 'Number of bytes evicted from the uncompressed cache.', `ProfileEvent_MMappedFileCacheHits` UInt64 COMMENT 'Number of times a file has been found in the MMap cache (for the \'mmap\' read_method), so we didn\'t have to mmap it again.', `ProfileEvent_MMappedFileCacheMisses` UInt64 COMMENT 'Number of times a file has not been found in the MMap cache (for the \'mmap\' read_method), so we had to mmap it again.', `ProfileEvent_OpenedFileCacheHits` UInt64 COMMENT 'Number of times a file has been found in the opened file cache, so we didn\'t have to open it again.', `ProfileEvent_OpenedFileCacheMisses` UInt64 COMMENT 'Number of times a file has been found in the opened file cache, so we had to open it again.', `ProfileEvent_OpenedFileCacheMicroseconds` UInt64 COMMENT 'Amount of time spent executing OpenedFileCache methods.', `ProfileEvent_AIOWrite` UInt64 COMMENT 'Number of writes with Linux or FreeBSD AIO interface', `ProfileEvent_AIOWriteBytes` UInt64 COMMENT 'Number of bytes written with Linux or FreeBSD AIO interface', `ProfileEvent_AIORead` UInt64 COMMENT 'Number of reads with Linux or FreeBSD AIO interface', `ProfileEvent_AIOReadBytes` UInt64 COMMENT 'Number of bytes read with Linux or FreeBSD AIO interface', `ProfileEvent_IOBufferAllocs` UInt64 COMMENT 'Number of allocations of IO buffers (for ReadBuffer/WriteBuffer).', `ProfileEvent_IOBufferAllocBytes` UInt64 COMMENT 'Number of bytes allocated for IO buffers (for ReadBuffer/WriteBuffer).', `ProfileEvent_ArenaAllocChunks` UInt64 COMMENT 'Number of chunks allocated for memory Arena (used for GROUP BY and similar operations)', `ProfileEvent_ArenaAllocBytes` UInt64 COMMENT 'Number of bytes allocated for memory Arena (used for GROUP BY and similar operations)', `ProfileEvent_FunctionExecute` UInt64 COMMENT 'Number of SQL ordinary function calls (SQL functions are called on per-block basis, so this number represents the number of blocks).', `ProfileEvent_TableFunctionExecute` UInt64 COMMENT 'Number of table function calls.', `ProfileEvent_DefaultImplementationForNullsRows` UInt64 COMMENT 'Number of rows processed by default implementation for nulls in function execution', `ProfileEvent_DefaultImplementationForNullsRowsWithNulls` UInt64 COMMENT 'Number of rows which contain null values processed by default implementation for nulls in function execution', `ProfileEvent_MarkCacheHits` UInt64 COMMENT 'Number of times an entry has been found in the mark cache, so we didn\'t have to load a mark file.', `ProfileEvent_MarkCacheMisses` UInt64 COMMENT 'Number of times an entry has not been found in the mark cache, so we had to load a mark file in memory, which is a costly operation, adding to query latency.', `ProfileEvent_PrimaryIndexCacheHits` UInt64 COMMENT 'Number of times an entry has been found in the primary index cache, so we didn\'t have to load a index file.', `ProfileEvent_PrimaryIndexCacheMisses` UInt64 COMMENT 'Number of times an entry has not been found in the primary index cache, so we had to load a index file in memory, which is a costly operation, adding to query latency.', `ProfileEvent_SkippingIndexCacheHits` UInt64 COMMENT 'Number of times an index granule has been found in the skipping index cache.', `ProfileEvent_SkippingIndexCacheMisses` UInt64 COMMENT 'Number of times an index granule has not been found in the skipping index cache and had to be read from disk.', `ProfileEvent_SkippingIndexCacheWeightLost` UInt64 COMMENT 'Approximate number of bytes evicted from the secondary index cache.', `ProfileEvent_QueryCacheHits` UInt64 COMMENT 'Number of times a query result has been found in the query cache (and query computation was avoided). Only updated for SELECT queries with SETTING use_query_cache = 1.', `ProfileEvent_QueryCacheMisses` UInt64 COMMENT 'Number of times a query result has not been found in the query cache (and required query computation). Only updated for SELECT queries with SETTING use_query_cache = 1.', `ProfileEvent_PageCacheChunkMisses` UInt64 COMMENT 'Number of times a chunk has not been found in the userspace page cache.', `ProfileEvent_PageCacheChunkShared` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, already in use by another thread.', `ProfileEvent_PageCacheChunkDataHits` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, with all pages intact.', `ProfileEvent_PageCacheChunkDataPartialHits` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, but some of its pages were evicted by the OS.', `ProfileEvent_PageCacheChunkDataMisses` UInt64 COMMENT 'Number of times a chunk has been found in the userspace page cache, not in use, but all its pages were evicted by the OS.', `ProfileEvent_PageCacheBytesUnpinnedRoundedToPages` UInt64 COMMENT 'Total size of populated pages in chunks that became evictable in PageCache. Rounded up to whole pages.', `ProfileEvent_PageCacheBytesUnpinnedRoundedToHugePages` UInt64 COMMENT 'See PageCacheBytesUnpinnedRoundedToPages, but rounded to huge pages. Use the ratio between the two as a measure of memory waste from using huge pages.', `ProfileEvent_CreatedReadBufferOrdinary` UInt64 COMMENT 'Number of times ordinary read buffer was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferDirectIO` UInt64 COMMENT 'Number of times a read buffer with O_DIRECT was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferDirectIOFailed` UInt64 COMMENT 'Number of times a read buffer with O_DIRECT was attempted to be created for reading data (while choosing among other read methods), but the OS did not allow it (due to lack of filesystem support or other reasons) and we fallen back to the ordinary reading method.', `ProfileEvent_CreatedReadBufferMMap` UInt64 COMMENT 'Number of times a read buffer using \'mmap\' was created for reading data (while choosing among other read methods).', `ProfileEvent_CreatedReadBufferMMapFailed` UInt64 COMMENT 'Number of times a read buffer with \'mmap\' was attempted to be created for reading data (while choosing among other read methods), but the OS did not allow it (due to lack of filesystem support or other reasons) and we fallen back to the ordinary reading method.', `ProfileEvent_DiskReadElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for read syscall. This include reads from page cache.', `ProfileEvent_DiskWriteElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for write syscall. This include writes to page cache.', `ProfileEvent_NetworkReceiveElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for data to receive or receiving data from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkSendElapsedMicroseconds` UInt64 COMMENT 'Total time spent waiting for data to send to network or sending data to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkReceiveBytes` UInt64 COMMENT 'Total number of bytes received from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_NetworkSendBytes` UInt64 COMMENT 'Total number of bytes send to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `ProfileEvent_GlobalThreadPoolExpansions` UInt64 COMMENT 'Counts the total number of times new threads have been added to the global thread pool. This metric indicates the frequency of expansions in the global thread pool to accommodate increased processing demands.', `ProfileEvent_GlobalThreadPoolShrinks` UInt64 COMMENT 'Counts the total number of times the global thread pool has shrunk by removing threads. This occurs when the number of idle threads exceeds max_thread_pool_free_size, indicating adjustments in the global thread pool size in response to decreased thread utilization.', `ProfileEvent_GlobalThreadPoolThreadCreationMicroseconds` UInt64 COMMENT 'Total time spent waiting for new threads to start.', `ProfileEvent_GlobalThreadPoolLockWaitMicroseconds` UInt64 COMMENT 'Total time threads have spent waiting for locks in the global thread pool.', `ProfileEvent_GlobalThreadPoolJobs` UInt64 COMMENT 'Counts the number of jobs that have been pushed to the global thread pool.', `ProfileEvent_GlobalThreadPoolJobWaitTimeMicroseconds` UInt64 COMMENT 'Measures the elapsed time from when a job is scheduled in the thread pool to when it is picked up for execution by a worker thread. This metric helps identify delays in job processing, indicating the responsiveness of the thread pool to new tasks.', `ProfileEvent_LocalThreadPoolExpansions` UInt64 COMMENT 'Counts the total number of times threads have been borrowed from the global thread pool to expand local thread pools.', `ProfileEvent_LocalThreadPoolShrinks` UInt64 COMMENT 'Counts the total number of times threads have been returned to the global thread pool from local thread pools.', `ProfileEvent_LocalThreadPoolThreadCreationMicroseconds` UInt64 COMMENT 'Total time local thread pools have spent waiting to borrow a thread from the global pool.', `ProfileEvent_LocalThreadPoolLockWaitMicroseconds` UInt64 COMMENT 'Total time threads have spent waiting for locks in the local thread pools.', `ProfileEvent_LocalThreadPoolJobs` UInt64 COMMENT 'Counts the number of jobs that have been pushed to the local thread pools.', `ProfileEvent_LocalThreadPoolBusyMicroseconds` UInt64 COMMENT 'Total time threads have spent executing the actual work.', `ProfileEvent_LocalThreadPoolJobWaitTimeMicroseconds` UInt64 COMMENT 'Measures the elapsed time from when a job is scheduled in the thread pool to when it is picked up for execution by a worker thread. This metric helps identify delays in job processing, indicating the responsiveness of the thread pool to new tasks.', `ProfileEvent_DiskS3GetRequestThrottlerCount` UInt64 COMMENT 'Number of DiskS3 GET and SELECT requests passed through throttler.', `ProfileEvent_DiskS3GetRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform DiskS3 GET and SELECT request throttling.', `ProfileEvent_DiskS3PutRequestThrottlerCount` UInt64 COMMENT 'Number of DiskS3 PUT, COPY, POST and LIST requests passed through throttler.', `ProfileEvent_DiskS3PutRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform DiskS3 PUT, COPY, POST and LIST request throttling.', `ProfileEvent_S3GetRequestThrottlerCount` UInt64 COMMENT 'Number of S3 GET and SELECT requests passed through throttler.', `ProfileEvent_S3GetRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform S3 GET and SELECT request throttling.', `ProfileEvent_S3PutRequestThrottlerCount` UInt64 COMMENT 'Number of S3 PUT, COPY, POST and LIST requests passed through throttler.', `ProfileEvent_S3PutRequestThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform S3 PUT, COPY, POST and LIST request throttling.', `ProfileEvent_RemoteReadThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_remote_read_network_bandwidth_for_server\'/\'max_remote_read_network_bandwidth\' throttler.', `ProfileEvent_RemoteReadThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_remote_read_network_bandwidth_for_server\'/\'max_remote_read_network_bandwidth\' throttling.', `ProfileEvent_RemoteWriteThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_remote_write_network_bandwidth_for_server\'/\'max_remote_write_network_bandwidth\' throttler.', `ProfileEvent_RemoteWriteThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_remote_write_network_bandwidth_for_server\'/\'max_remote_write_network_bandwidth\' throttling.', `ProfileEvent_LocalReadThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_local_read_bandwidth_for_server\'/\'max_local_read_bandwidth\' throttler.', `ProfileEvent_LocalReadThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_local_read_bandwidth_for_server\'/\'max_local_read_bandwidth\' throttling.', `ProfileEvent_LocalWriteThrottlerBytes` UInt64 COMMENT 'Bytes passed through \'max_local_write_bandwidth_for_server\'/\'max_local_write_bandwidth\' throttler.', `ProfileEvent_LocalWriteThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform \'max_local_write_bandwidth_for_server\'/\'max_local_write_bandwidth\' throttling.', `ProfileEvent_ThrottlerSleepMicroseconds` UInt64 COMMENT 'Total time a query was sleeping to conform all throttling settings.', `ProfileEvent_ReadTasksWithAppliedMutationsOnFly` UInt64 COMMENT 'Total number of parts for which there was any mutation applied on fly', `ProfileEvent_MutationsAppliedOnFlyInAllReadTasks` UInt64 COMMENT 'The sum of number of applied mutations on-fly for part among all read parts', `ProfileEvent_SchedulerIOReadRequests` UInt64 COMMENT 'Resource requests passed through scheduler for IO reads.', `ProfileEvent_SchedulerIOReadBytes` UInt64 COMMENT 'Bytes passed through scheduler for IO reads.', `ProfileEvent_SchedulerIOReadWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting on resource requests for IO reads.', `ProfileEvent_SchedulerIOWriteRequests` UInt64 COMMENT 'Resource requests passed through scheduler for IO writes.', `ProfileEvent_SchedulerIOWriteBytes` UInt64 COMMENT 'Bytes passed through scheduler for IO writes.', `ProfileEvent_SchedulerIOWriteWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting on resource requests for IO writes.', `ProfileEvent_QueryMaskingRulesMatch` UInt64 COMMENT 'Number of times query masking rules was successfully matched.', `ProfileEvent_ReplicatedPartFetches` UInt64 COMMENT 'Number of times a data part was downloaded from replica of a ReplicatedMergeTree table.', `ProfileEvent_ReplicatedPartFailedFetches` UInt64 COMMENT 'Number of times a data part was failed to download from replica of a ReplicatedMergeTree table.', `ProfileEvent_ObsoleteReplicatedParts` UInt64 COMMENT 'Number of times a data part was covered by another data part that has been fetched from a replica (so, we have marked a covered data part as obsolete and no longer needed).', `ProfileEvent_ReplicatedPartMerges` UInt64 COMMENT 'Number of times data parts of ReplicatedMergeTree tables were successfully merged.', `ProfileEvent_ReplicatedPartFetchesOfMerged` UInt64 COMMENT 'Number of times we prefer to download already merged part from replica of ReplicatedMergeTree table instead of performing a merge ourself (usually we prefer doing a merge ourself to save network traffic). This happens when we have not all source parts to perform a merge or when the data part is old enough.', `ProfileEvent_ReplicatedPartMutations` UInt64 COMMENT 'Number of times data parts of ReplicatedMergeTree tables were successfully mutated.', `ProfileEvent_ReplicatedPartChecks` UInt64 COMMENT 'Number of times we had to perform advanced search for a data part on replicas or to clarify the need of an existing data part.', `ProfileEvent_ReplicatedPartChecksFailed` UInt64 COMMENT 'Number of times the advanced search for a data part on replicas did not give result or when unexpected part has been found and moved away.', `ProfileEvent_ReplicatedDataLoss` UInt64 COMMENT 'Number of times a data part that we wanted doesn\'t exist on any replica (even on replicas that are offline right now). That data parts are definitely lost. This is normal due to asynchronous replication (if quorum inserts were not enabled), when the replica on which the data part was written was failed and when it became online after fail it doesn\'t contain that data part.', `ProfileEvent_ReplicatedCoveredPartsInZooKeeperOnStart` UInt64 COMMENT 'For debugging purposes. Number of parts in ZooKeeper that have a covering part, but doesn\'t exist on disk. Checked on server start.', `ProfileEvent_InsertedRows` UInt64 COMMENT 'Number of rows INSERTed to all tables.', `ProfileEvent_InsertedBytes` UInt64 COMMENT 'Number of bytes (uncompressed; for columns as they stored in memory) INSERTed to all tables.', `ProfileEvent_DelayedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a MergeTree table was throttled due to high number of active data parts for partition.', `ProfileEvent_RejectedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a MergeTree table was rejected with \'Too many parts\' exception due to high number of active data parts for partition.', `ProfileEvent_DelayedInsertsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the INSERT of a block to a MergeTree table was throttled due to high number of active data parts for partition.', `ProfileEvent_DelayedMutations` UInt64 COMMENT 'Number of times the mutation of a MergeTree table was throttled due to high number of unfinished mutations for table.', `ProfileEvent_RejectedMutations` UInt64 COMMENT 'Number of times the mutation of a MergeTree table was rejected with \'Too many mutations\' exception due to high number of unfinished mutations for table.', `ProfileEvent_DelayedMutationsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the mutation of a MergeTree table was throttled due to high number of unfinished mutations for table.', `ProfileEvent_DistributedDelayedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a Distributed table was throttled due to high number of pending bytes.', `ProfileEvent_DistributedRejectedInserts` UInt64 COMMENT 'Number of times the INSERT of a block to a Distributed table was rejected with \'Too many bytes\' exception due to high number of pending bytes.', `ProfileEvent_DistributedDelayedInsertsMilliseconds` UInt64 COMMENT 'Total number of milliseconds spent while the INSERT of a block to a Distributed table was throttled due to high number of pending bytes.', `ProfileEvent_DuplicatedInsertedBlocks` UInt64 COMMENT 'Number of times the INSERTed block to a ReplicatedMergeTree table was deduplicated.', `ProfileEvent_ZooKeeperInit` UInt64 COMMENT 'Number of times connection with ZooKeeper has been established.', `ProfileEvent_ZooKeeperTransactions` UInt64 COMMENT 'Number of ZooKeeper operations, which include both read and write operations as well as multi-transactions.', `ProfileEvent_ZooKeeperList` UInt64 COMMENT 'Number of \'list\' (getChildren) requests to ZooKeeper.', `ProfileEvent_ZooKeeperCreate` UInt64 COMMENT 'Number of \'create\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperRemove` UInt64 COMMENT 'Number of \'remove\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperExists` UInt64 COMMENT 'Number of \'exists\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperGet` UInt64 COMMENT 'Number of \'get\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperSet` UInt64 COMMENT 'Number of \'set\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperMulti` UInt64 COMMENT 'Number of \'multi\' requests to ZooKeeper (compound transactions).', `ProfileEvent_ZooKeeperCheck` UInt64 COMMENT 'Number of \'check\' requests to ZooKeeper. Usually they don\'t make sense in isolation, only as part of a complex transaction.', `ProfileEvent_ZooKeeperSync` UInt64 COMMENT 'Number of \'sync\' requests to ZooKeeper. These requests are rarely needed or usable.', `ProfileEvent_ZooKeeperReconfig` UInt64 COMMENT 'Number of \'reconfig\' requests to ZooKeeper.', `ProfileEvent_ZooKeeperClose` UInt64 COMMENT 'Number of times connection with ZooKeeper has been closed voluntary.', `ProfileEvent_ZooKeeperWatchResponse` UInt64 COMMENT 'Number of times watch notification has been received from ZooKeeper.', `ProfileEvent_ZooKeeperUserExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper related to the data (no node, bad version or similar).', `ProfileEvent_ZooKeeperHardwareExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper related to network (connection loss or similar).', `ProfileEvent_ZooKeeperOtherExceptions` UInt64 COMMENT 'Number of exceptions while working with ZooKeeper other than ZooKeeperUserExceptions and ZooKeeperHardwareExceptions.', `ProfileEvent_ZooKeeperWaitMicroseconds` UInt64 COMMENT 'Number of microseconds spent waiting for responses from ZooKeeper after creating a request, summed across all the requesting threads.', `ProfileEvent_ZooKeeperBytesSent` UInt64 COMMENT 'Number of bytes send over network while communicating with ZooKeeper.', `ProfileEvent_ZooKeeperBytesReceived` UInt64 COMMENT 'Number of bytes received over network while communicating with ZooKeeper.', `ProfileEvent_DistributedConnectionTries` UInt64 COMMENT 'Total count of distributed connection attempts.', `ProfileEvent_DistributedConnectionUsable` UInt64 COMMENT 'Total count of successful distributed connections to a usable server (with required table, but maybe stale).', `ProfileEvent_DistributedConnectionFailTry` UInt64 COMMENT 'Total count when distributed connection fails with retry.', `ProfileEvent_DistributedConnectionMissingTable` UInt64 COMMENT 'Number of times we rejected a replica from a distributed query, because it did not contain a table needed for the query.', `ProfileEvent_DistributedConnectionStaleReplica` UInt64 COMMENT 'Number of times we rejected a replica from a distributed query, because some table needed for a query had replication lag higher than the configured threshold.', `ProfileEvent_DistributedConnectionSkipReadOnlyReplica` UInt64 COMMENT 'Number of replicas skipped during INSERT into Distributed table due to replicas being read-only', `ProfileEvent_DistributedConnectionFailAtAll` UInt64 COMMENT 'Total count when distributed connection fails after all retries finished.', `ProfileEvent_HedgedRequestsChangeReplica` UInt64 COMMENT 'Total count when timeout for changing replica expired in hedged requests.', `ProfileEvent_SuspendSendingQueryToShard` UInt64 COMMENT 'Total count when sending query to shard was suspended when async_query_sending_for_remote is enabled.', `ProfileEvent_CompileFunction` UInt64 COMMENT 'Number of times a compilation of generated LLVM code (to create fused function for complex expressions) was initiated.', `ProfileEvent_CompiledFunctionExecute` UInt64 COMMENT 'Number of times a compiled function was executed.', `ProfileEvent_CompileExpressionsMicroseconds` UInt64 COMMENT 'Total time spent for compilation of expressions to LLVM code.', `ProfileEvent_CompileExpressionsBytes` UInt64 COMMENT 'Number of bytes used for expressions compilation.', `ProfileEvent_ExecuteShellCommand` UInt64 COMMENT 'Number of shell command executions.', `ProfileEvent_ExternalProcessingCompressedBytesTotal` UInt64 COMMENT 'Number of compressed bytes written by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalProcessingUncompressedBytesTotal` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalProcessingFilesTotal` UInt64 COMMENT 'Number of files used by external processing (sorting/aggragating/joining)', `ProfileEvent_ExternalSortWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for sorting in external memory.', `ProfileEvent_ExternalSortMerge` UInt64 COMMENT 'Number of times temporary files were merged for sorting in external memory.', `ProfileEvent_ExternalSortCompressedBytes` UInt64 COMMENT 'Number of compressed bytes written for sorting in external memory.', `ProfileEvent_ExternalSortUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written for sorting in external memory.', `ProfileEvent_ExternalAggregationWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for aggregation in external memory.', `ProfileEvent_ExternalAggregationMerge` UInt64 COMMENT 'Number of times temporary files were merged for aggregation in external memory.', `ProfileEvent_ExternalAggregationCompressedBytes` UInt64 COMMENT 'Number of bytes written to disk for aggregation in external memory.', `ProfileEvent_ExternalAggregationUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written to disk for aggregation in external memory.', `ProfileEvent_ExternalJoinWritePart` UInt64 COMMENT 'Number of times a temporary file was written to disk for JOIN in external memory.', `ProfileEvent_ExternalJoinMerge` UInt64 COMMENT 'Number of times temporary files were merged for JOIN in external memory.', `ProfileEvent_ExternalJoinCompressedBytes` UInt64 COMMENT 'Number of compressed bytes written for JOIN in external memory.', `ProfileEvent_ExternalJoinUncompressedBytes` UInt64 COMMENT 'Amount of data (uncompressed, before compression) written for JOIN in external memory.', `ProfileEvent_IcebergPartitionPrunnedFiles` UInt64 COMMENT 'Number of skipped files during Iceberg partition pruning', `ProfileEvent_JoinBuildTableRowCount` UInt64 COMMENT 'Total number of rows in the build table for a JOIN operation.', `ProfileEvent_JoinProbeTableRowCount` UInt64 COMMENT 'Total number of rows in the probe table for a JOIN operation.', `ProfileEvent_JoinResultRowCount` UInt64 COMMENT 'Total number of rows in the result of a JOIN operation.', `ProfileEvent_SlowRead` UInt64 COMMENT 'Number of reads from a file that were slow. This indicate system overload. Thresholds are controlled by read_backoff_* settings.', `ProfileEvent_ReadBackoff` UInt64 COMMENT 'Number of times the number of query processing threads was lowered due to slow reads.', `ProfileEvent_ReplicaPartialShutdown` UInt64 COMMENT 'How many times Replicated table has to deinitialize its state due to session expiration in ZooKeeper. The state is reinitialized every time when ZooKeeper is available again.', `ProfileEvent_SelectedParts` UInt64 COMMENT 'Number of data parts selected to read from a MergeTree table.', `ProfileEvent_SelectedPartsTotal` UInt64 COMMENT 'Number of total data parts before selecting which ones to read from a MergeTree table.', `ProfileEvent_SelectedRanges` UInt64 COMMENT 'Number of (non-adjacent) ranges in all data parts selected to read from a MergeTree table.', `ProfileEvent_SelectedMarks` UInt64 COMMENT 'Number of marks (index granules) selected to read from a MergeTree table.', `ProfileEvent_SelectedMarksTotal` UInt64 COMMENT 'Number of total marks (index granules) before selecting which ones to read from a MergeTree table.', `ProfileEvent_SelectedRows` UInt64 COMMENT 'Number of rows SELECTed from all tables.', `ProfileEvent_SelectedBytes` UInt64 COMMENT 'Number of bytes (uncompressed; for columns as they stored in memory) SELECTed from all tables.', `ProfileEvent_RowsReadByMainReader` UInt64 COMMENT 'Number of rows read from MergeTree tables by the main reader (after PREWHERE step).', `ProfileEvent_RowsReadByPrewhereReaders` UInt64 COMMENT 'Number of rows read from MergeTree tables (in total) by prewhere readers.', `ProfileEvent_LoadedDataParts` UInt64 COMMENT 'Number of data parts loaded by MergeTree tables during initialization.', `ProfileEvent_LoadedDataPartsMicroseconds` UInt64 COMMENT 'Microseconds spent by MergeTree tables for loading data parts during initialization.', `ProfileEvent_WaitMarksLoadMicroseconds` UInt64 COMMENT 'Time spent loading marks', `ProfileEvent_BackgroundLoadingMarksTasks` UInt64 COMMENT 'Number of background tasks for loading marks', `ProfileEvent_LoadingMarksTasksCanceled` UInt64 COMMENT 'Number of times background tasks for loading marks were canceled', `ProfileEvent_LoadedMarksFiles` UInt64 COMMENT 'Number of mark files loaded.', `ProfileEvent_LoadedMarksCount` UInt64 COMMENT 'Number of marks loaded (total across columns).', `ProfileEvent_LoadedMarksMemoryBytes` UInt64 COMMENT 'Size of in-memory representations of loaded marks.', `ProfileEvent_LoadedPrimaryIndexFiles` UInt64 COMMENT 'Number of primary index files loaded.', `ProfileEvent_LoadedPrimaryIndexRows` UInt64 COMMENT 'Number of rows of primary key loaded.', `ProfileEvent_LoadedPrimaryIndexBytes` UInt64 COMMENT 'Number of rows of primary key loaded.', `ProfileEvent_Merge` UInt64 COMMENT 'Number of launched background merges.', `ProfileEvent_MergeSourceParts` UInt64 COMMENT 'Number of source parts scheduled for merges.', `ProfileEvent_MergedRows` UInt64 COMMENT 'Rows read for background merges. This is the number of rows before merge.', `ProfileEvent_MergedColumns` UInt64 COMMENT 'Number of columns merged during the horizontal stage of merges.', `ProfileEvent_GatheredColumns` UInt64 COMMENT 'Number of columns gathered during the vertical stage of merges.', `ProfileEvent_MergedUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) that was read for background merges. This is the number before merge.', `ProfileEvent_MergeTotalMilliseconds` UInt64 COMMENT 'Total time spent for background merges', `ProfileEvent_MergeExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of background merges', `ProfileEvent_MergeHorizontalStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for horizontal stage of background merges', `ProfileEvent_MergeHorizontalStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of horizontal stage of background merges', `ProfileEvent_MergeVerticalStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for vertical stage of background merges', `ProfileEvent_MergeVerticalStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of vertical stage of background merges', `ProfileEvent_MergeProjectionStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for projection stage of background merges', `ProfileEvent_MergeProjectionStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of projection stage of background merges', `ProfileEvent_MergePrewarmStageTotalMilliseconds` UInt64 COMMENT 'Total time spent for prewarm stage of background merges', `ProfileEvent_MergePrewarmStageExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of prewarm stage of background merges', `ProfileEvent_MergingSortedMilliseconds` UInt64 COMMENT 'Total time spent while merging sorted columns', `ProfileEvent_AggregatingSortedMilliseconds` UInt64 COMMENT 'Total time spent while aggregating sorted columns', `ProfileEvent_CollapsingSortedMilliseconds` UInt64 COMMENT 'Total time spent while collapsing sorted columns', `ProfileEvent_ReplacingSortedMilliseconds` UInt64 COMMENT 'Total time spent while replacing sorted columns', `ProfileEvent_SummingSortedMilliseconds` UInt64 COMMENT 'Total time spent while summing sorted columns', `ProfileEvent_VersionedCollapsingSortedMilliseconds` UInt64 COMMENT 'Total time spent while version collapsing sorted columns', `ProfileEvent_GatheringColumnMilliseconds` UInt64 COMMENT 'Total time spent while gathering columns for vertical merge', `ProfileEvent_MutationTotalParts` UInt64 COMMENT 'Number of total parts for which mutations tried to be applied', `ProfileEvent_MutationUntouchedParts` UInt64 COMMENT 'Number of total parts for which mutations tried to be applied but which was completely skipped according to predicate', `ProfileEvent_MutatedRows` UInt64 COMMENT 'Rows read for mutations. This is the number of rows before mutation', `ProfileEvent_MutatedUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) that was read for mutations. This is the number before mutation.', `ProfileEvent_MutationTotalMilliseconds` UInt64 COMMENT 'Total time spent for mutations.', `ProfileEvent_MutationExecuteMilliseconds` UInt64 COMMENT 'Total busy time spent for execution of mutations.', `ProfileEvent_MutationAllPartColumns` UInt64 COMMENT 'Number of times when task to mutate all columns in part was created', `ProfileEvent_MutationSomePartColumns` UInt64 COMMENT 'Number of times when task to mutate some columns in part was created', `ProfileEvent_MutateTaskProjectionsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating projections in mutations', `ProfileEvent_MergeTreeDataWriterRows` UInt64 COMMENT 'Number of rows INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterCompressedBytes` UInt64 COMMENT 'Bytes written to filesystem for data INSERTed to MergeTree tables.', `ProfileEvent_MergeTreeDataWriterBlocks` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables. Each block forms a data part of level zero.', `ProfileEvent_MergeTreeDataWriterBlocksAlreadySorted` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables that appeared to be already sorted.', `ProfileEvent_MergeTreeDataWriterSkipIndicesCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating skip indices', `ProfileEvent_MergeTreeDataWriterStatisticsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating statistics', `ProfileEvent_MergeTreeDataWriterSortingBlocksMicroseconds` UInt64 COMMENT 'Time spent sorting blocks', `ProfileEvent_MergeTreeDataWriterMergingBlocksMicroseconds` UInt64 COMMENT 'Time spent merging input blocks (for special MergeTree engines)', `ProfileEvent_MergeTreeDataWriterProjectionsCalculationMicroseconds` UInt64 COMMENT 'Time spent calculating projections', `ProfileEvent_MergeTreeDataProjectionWriterSortingBlocksMicroseconds` UInt64 COMMENT 'Time spent sorting blocks (for projection it might be a key different from table\'s sorting key)', `ProfileEvent_MergeTreeDataProjectionWriterMergingBlocksMicroseconds` UInt64 COMMENT 'Time spent merging blocks', `ProfileEvent_InsertedWideParts` UInt64 COMMENT 'Number of parts inserted in Wide format.', `ProfileEvent_InsertedCompactParts` UInt64 COMMENT 'Number of parts inserted in Compact format.', `ProfileEvent_MergedIntoWideParts` UInt64 COMMENT 'Number of parts merged into Wide format.', `ProfileEvent_MergedIntoCompactParts` UInt64 COMMENT 'Number of parts merged into Compact format.', `ProfileEvent_MergeTreeDataProjectionWriterRows` UInt64 COMMENT 'Number of rows INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterUncompressedBytes` UInt64 COMMENT 'Uncompressed bytes (for columns as they stored in memory) INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterCompressedBytes` UInt64 COMMENT 'Bytes written to filesystem for data INSERTed to MergeTree tables projection.', `ProfileEvent_MergeTreeDataProjectionWriterBlocks` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables projection. Each block forms a data part of level zero.', `ProfileEvent_MergeTreeDataProjectionWriterBlocksAlreadySorted` UInt64 COMMENT 'Number of blocks INSERTed to MergeTree tables projection that appeared to be already sorted.', `ProfileEvent_CannotRemoveEphemeralNode` UInt64 COMMENT 'Number of times an error happened while trying to remove ephemeral node. This is not an issue, because our implementation of ZooKeeper library guarantee that the session will expire and the node will be removed.', `ProfileEvent_RegexpWithMultipleNeedlesCreated` UInt64 COMMENT 'Regular expressions with multiple needles (VectorScan library) compiled.', `ProfileEvent_RegexpWithMultipleNeedlesGlobalCacheHit` UInt64 COMMENT 'Number of times we fetched compiled regular expression with multiple needles (VectorScan library) from the global cache.', `ProfileEvent_RegexpWithMultipleNeedlesGlobalCacheMiss` UInt64 COMMENT 'Number of times we failed to fetch compiled regular expression with multiple needles (VectorScan library) from the global cache.', `ProfileEvent_RegexpLocalCacheHit` UInt64 COMMENT 'Number of times we fetched compiled regular expression from a local cache.', `ProfileEvent_RegexpLocalCacheMiss` UInt64 COMMENT 'Number of times we failed to fetch compiled regular expression from a local cache.', `ProfileEvent_ContextLock` UInt64 COMMENT 'Number of times the lock of Context was acquired or tried to acquire. This is global lock.', `ProfileEvent_ContextLockWaitMicroseconds` UInt64 COMMENT 'Context lock wait time in microseconds', `ProfileEvent_StorageBufferFlush` UInt64 COMMENT 'Number of times a buffer in a \'Buffer\' table was flushed.', `ProfileEvent_StorageBufferErrorOnFlush` UInt64 COMMENT 'Number of times a buffer in the \'Buffer\' table has not been able to flush due to error writing in the destination table.', `ProfileEvent_StorageBufferPassedAllMinThresholds` UInt64 COMMENT 'Number of times a criteria on min thresholds has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedTimeMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max time threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedRowsMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max rows threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedBytesMaxThreshold` UInt64 COMMENT 'Number of times a criteria on max bytes threshold has been reached to flush a buffer in a \'Buffer\' table.', `ProfileEvent_StorageBufferPassedTimeFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on time has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferPassedRowsFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on rows has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferPassedBytesFlushThreshold` UInt64 COMMENT 'Number of times background-only flush threshold on bytes has been reached to flush a buffer in a \'Buffer\' table. This is expert-only metric. If you read this and you are not an expert, stop reading.', `ProfileEvent_StorageBufferLayerLockReadersWaitMilliseconds` UInt64 COMMENT 'Time for waiting for Buffer layer during reading.', `ProfileEvent_StorageBufferLayerLockWritersWaitMilliseconds` UInt64 COMMENT 'Time for waiting free Buffer layer to write to (can be used to tune Buffer layers).', `ProfileEvent_SystemLogErrorOnFlush` UInt64 COMMENT 'Number of times any of the system logs have failed to flush to the corresponding system table. Attempts to flush are repeated.', `ProfileEvent_DictCacheKeysRequested` UInt64 COMMENT 'Number of keys requested from the data source for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheKeysRequestedMiss` UInt64 COMMENT 'Number of keys requested from the data source for dictionaries of \'cache\' types but not found in the data source.', `ProfileEvent_DictCacheKeysRequestedFound` UInt64 COMMENT 'Number of keys requested from the data source for dictionaries of \'cache\' types and found in the data source.', `ProfileEvent_DictCacheKeysExpired` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and found in the cache but they were obsolete.', `ProfileEvent_DictCacheKeysNotFound` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and not found.', `ProfileEvent_DictCacheKeysHit` UInt64 COMMENT 'Number of keys looked up in the dictionaries of \'cache\' types and found in the cache.', `ProfileEvent_DictCacheRequestTimeNs` UInt64 COMMENT 'Number of nanoseconds spend in querying the external data sources for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheRequests` UInt64 COMMENT 'Number of bulk requests to the external data sources for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheLockWriteNs` UInt64 COMMENT 'Number of nanoseconds spend in waiting for write lock to update the data for the dictionaries of \'cache\' types.', `ProfileEvent_DictCacheLockReadNs` UInt64 COMMENT 'Number of nanoseconds spend in waiting for read lock to lookup the data for the dictionaries of \'cache\' types.', `ProfileEvent_DistributedSyncInsertionTimeoutExceeded` UInt64 COMMENT 'A timeout has exceeded while waiting for shards during synchronous insertion into a Distributed table (with \'distributed_foreground_insert\' = 1)', `ProfileEvent_DistributedAsyncInsertionFailures` UInt64 COMMENT 'Number of failures for asynchronous insertion into a Distributed table (with \'distributed_foreground_insert\' = 0)', `ProfileEvent_DataAfterMergeDiffersFromReplica` UInt64 COMMENT '\nNumber of times data after merge is not byte-identical to the data on another replicas. There could be several reasons:\n1. Using newer version of compression library after server update.\n2. Using another compression method.\n3. Non-deterministic compression algorithm (highly unlikely).\n4. Non-deterministic merge algorithm due to logical error in code.\n5. Data corruption in memory due to bug in code.\n6. Data corruption in memory due to hardware issue.\n7. Manual modification of source data after server startup.\n8. Manual modification of checksums stored in ZooKeeper.\n9. Part format related settings like \'enable_mixed_granularity_parts\' are different on different replicas.\nThe server successfully detected this situation and will download merged part from the replica to force the byte-identical result.\n', `ProfileEvent_DataAfterMutationDiffersFromReplica` UInt64 COMMENT 'Number of times data after mutation is not byte-identical to the data on other replicas. In addition to the reasons described in \'DataAfterMergeDiffersFromReplica\', it is also possible due to non-deterministic mutation.', `ProfileEvent_PolygonsAddedToPool` UInt64 COMMENT 'A polygon has been added to the cache (pool) for the \'pointInPolygon\' function.', `ProfileEvent_PolygonsInPoolAllocatedBytes` UInt64 COMMENT 'The number of bytes for polygons added to the cache (pool) for the \'pointInPolygon\' function.', `ProfileEvent_USearchAddCount` UInt64 COMMENT 'Number of vectors added to usearch indexes.', `ProfileEvent_USearchAddVisitedMembers` UInt64 COMMENT 'Number of nodes visited when adding vectors to usearch indexes.', `ProfileEvent_USearchAddComputedDistances` UInt64 COMMENT 'Number of times distance was computed when adding vectors to usearch indexes.', `ProfileEvent_USearchSearchCount` UInt64 COMMENT 'Number of search operations performed in usearch indexes.', `ProfileEvent_USearchSearchVisitedMembers` UInt64 COMMENT 'Number of nodes visited when searching in usearch indexes.', `ProfileEvent_USearchSearchComputedDistances` UInt64 COMMENT 'Number of times distance was computed when searching usearch indexes.', `ProfileEvent_RWLockAcquiredReadLocks` UInt64 COMMENT 'Number of times a read lock was acquired (in a heavy RWLock).', `ProfileEvent_RWLockAcquiredWriteLocks` UInt64 COMMENT 'Number of times a write lock was acquired (in a heavy RWLock).', `ProfileEvent_RWLockReadersWaitMilliseconds` UInt64 COMMENT 'Total time spent waiting for a read lock to be acquired (in a heavy RWLock).', `ProfileEvent_RWLockWritersWaitMilliseconds` UInt64 COMMENT 'Total time spent waiting for a write lock to be acquired (in a heavy RWLock).', `ProfileEvent_DNSError` UInt64 COMMENT 'Total count of errors in DNS resolution', `ProfileEvent_PartsLockHoldMicroseconds` UInt64 COMMENT 'Total time spent holding data parts lock in MergeTree tables', `ProfileEvent_PartsLockWaitMicroseconds` UInt64 COMMENT 'Total time spent waiting for data parts lock in MergeTree tables', `ProfileEvent_RealTimeMicroseconds` UInt64 COMMENT 'Total (wall clock) time spent in processing (queries and other tasks) threads (note that this is a sum).', `ProfileEvent_UserTimeMicroseconds` UInt64 COMMENT 'Total time spent in processing (queries and other tasks) threads executing CPU instructions in user mode. This includes time CPU pipeline was stalled due to main memory access, cache misses, branch mispredictions, hyper-threading, etc.', `ProfileEvent_SystemTimeMicroseconds` UInt64 COMMENT 'Total time spent in processing (queries and other tasks) threads executing CPU instructions in OS kernel mode. This is time spent in syscalls, excluding waiting time during blocking syscalls.', `ProfileEvent_MemoryOvercommitWaitTimeMicroseconds` UInt64 COMMENT 'Total time spent in waiting for memory to be freed in OvercommitTracker.', `ProfileEvent_MemoryAllocatorPurge` UInt64 COMMENT 'Total number of times memory allocator purge was requested', `ProfileEvent_MemoryAllocatorPurgeTimeMicroseconds` UInt64 COMMENT 'Total number of times memory allocator purge was requested', `ProfileEvent_SoftPageFaults` UInt64 COMMENT 'The number of soft page faults in query execution threads. Soft page fault usually means a miss in the memory allocator cache, which requires a new memory mapping from the OS and subsequent allocation of a page of physical memory.', `ProfileEvent_HardPageFaults` UInt64 COMMENT 'The number of hard page faults in query execution threads. High values indicate either that you forgot to turn off swap on your server, or eviction of memory pages of the ClickHouse binary during very high memory pressure, or successful usage of the \'mmap\' read method for the tables data.', `ProfileEvent_OSIOWaitMicroseconds` UInt64 COMMENT 'Total time a thread spent waiting for a result of IO operation, from the OS point of view. This is real IO that doesn\'t include page cache.', `ProfileEvent_OSCPUWaitMicroseconds` UInt64 COMMENT 'Total time a thread was ready for execution but waiting to be scheduled by OS, from the OS point of view.', `ProfileEvent_OSCPUVirtualTimeMicroseconds` UInt64 COMMENT 'CPU time spent seen by OS. Does not include involuntary waits due to virtualization.', `ProfileEvent_OSReadBytes` UInt64 COMMENT 'Number of bytes read from disks or block devices. Doesn\'t include bytes read from page cache. May include excessive data due to block size, readahead, etc.', `ProfileEvent_OSWriteBytes` UInt64 COMMENT 'Number of bytes written to disks or block devices. Doesn\'t include bytes that are in page cache dirty pages. May not include data that was written by OS asynchronously.', `ProfileEvent_OSReadChars` UInt64 COMMENT 'Number of bytes read from filesystem, including page cache.', `ProfileEvent_OSWriteChars` UInt64 COMMENT 'Number of bytes written to filesystem, including page cache.', `ProfileEvent_ParallelReplicasHandleRequestMicroseconds` UInt64 COMMENT 'Time spent processing requests for marks from replicas', `ProfileEvent_ParallelReplicasHandleAnnouncementMicroseconds` UInt64 COMMENT 'Time spent processing replicas announcements', `ProfileEvent_ParallelReplicasAnnouncementMicroseconds` UInt64 COMMENT 'Time spent to send an announcement', `ProfileEvent_ParallelReplicasReadRequestMicroseconds` UInt64 COMMENT 'Time spent for read requests', `ProfileEvent_ParallelReplicasReadAssignedMarks` UInt64 COMMENT 'Sum across all replicas of how many of scheduled marks were assigned by consistent hash', `ProfileEvent_ParallelReplicasReadUnassignedMarks` UInt64 COMMENT 'Sum across all replicas of how many unassigned marks were scheduled', `ProfileEvent_ParallelReplicasReadAssignedForStealingMarks` UInt64 COMMENT 'Sum across all replicas of how many of scheduled marks were assigned for stealing by consistent hash', `ProfileEvent_ParallelReplicasReadMarks` UInt64 COMMENT 'How many marks were read by the given replica', `ProfileEvent_ParallelReplicasStealingByHashMicroseconds` UInt64 COMMENT 'Time spent collecting segments meant for stealing by hash', `ProfileEvent_ParallelReplicasProcessingPartsMicroseconds` UInt64 COMMENT 'Time spent processing data parts', `ProfileEvent_ParallelReplicasStealingLeftoversMicroseconds` UInt64 COMMENT 'Time spent collecting orphaned segments', `ProfileEvent_ParallelReplicasCollectingOwnedSegmentsMicroseconds` UInt64 COMMENT 'Time spent collecting segments meant by hash', `ProfileEvent_ParallelReplicasNumRequests` UInt64 COMMENT 'Number of requests to the initiator.', `ProfileEvent_ParallelReplicasDeniedRequests` UInt64 COMMENT 'Number of completely denied requests to the initiator', `ProfileEvent_CacheWarmerBytesDownloaded` UInt64 COMMENT 'Amount of data fetched into filesystem cache by dedicated background threads.', `ProfileEvent_CacheWarmerDataPartsDownloaded` UInt64 COMMENT 'Number of data parts that were fully fetched by CacheWarmer.', `ProfileEvent_IgnoredColdParts` UInt64 COMMENT 'See setting ignore_cold_parts_seconds. Number of times read queries ignored very new parts that weren\'t pulled into cache by CacheWarmer yet.', `ProfileEvent_PreferredWarmedUnmergedParts` UInt64 COMMENT 'See setting prefer_warmed_unmerged_parts_seconds. Number of times read queries used outdated pre-merge parts that are in cache instead of merged part that wasn\'t pulled into cache by CacheWarmer yet.', `ProfileEvent_PerfCPUCycles` UInt64 COMMENT 'Total cycles. Be wary of what happens during CPU frequency scaling.', `ProfileEvent_PerfInstructions` UInt64 COMMENT 'Retired instructions. Be careful, these can be affected by various issues, most notably hardware interrupt counts.', `ProfileEvent_PerfCacheReferences` UInt64 COMMENT 'Cache accesses. Usually, this indicates Last Level Cache accesses, but this may vary depending on your CPU. This may include prefetches and coherency messages; again this depends on the design of your CPU.', `ProfileEvent_PerfCacheMisses` UInt64 COMMENT 'Cache misses. Usually this indicates Last Level Cache misses; this is intended to be used in conjunction with the PERFCOUNTHWCACHEREFERENCES event to calculate cache miss rates.', `ProfileEvent_PerfBranchInstructions` UInt64 COMMENT 'Retired branch instructions. Prior to Linux 2.6.35, this used the wrong event on AMD processors.', `ProfileEvent_PerfBranchMisses` UInt64 COMMENT 'Mispredicted branch instructions.', `ProfileEvent_PerfBusCycles` UInt64 COMMENT 'Bus cycles, which can be different from total cycles.', `ProfileEvent_PerfStalledCyclesFrontend` UInt64 COMMENT 'Stalled cycles during issue.', `ProfileEvent_PerfStalledCyclesBackend` UInt64 COMMENT 'Stalled cycles during retirement.', `ProfileEvent_PerfRefCPUCycles` UInt64 COMMENT 'Total cycles; not affected by CPU frequency scaling.', `ProfileEvent_PerfCPUClock` UInt64 COMMENT 'The CPU clock, a high-resolution per-CPU timer', `ProfileEvent_PerfTaskClock` UInt64 COMMENT 'A clock count specific to the task that is running', `ProfileEvent_PerfContextSwitches` UInt64 COMMENT 'Number of context switches', `ProfileEvent_PerfCPUMigrations` UInt64 COMMENT 'Number of times the process has migrated to a new CPU', `ProfileEvent_PerfAlignmentFaults` UInt64 COMMENT 'Number of alignment faults. These happen when unaligned memory accesses happen; the kernel can handle these but it reduces performance. This happens only on some architectures (never on x86).', `ProfileEvent_PerfEmulationFaults` UInt64 COMMENT 'Number of emulation faults. The kernel sometimes traps on unimplemented instructions and emulates them for user space. This can negatively impact performance.', `ProfileEvent_PerfMinEnabledTime` UInt64 COMMENT 'For all events, minimum time that an event was enabled. Used to track event multiplexing influence', `ProfileEvent_PerfMinEnabledRunningTime` UInt64 COMMENT 'Running time for event with minimum enabled time. Used to track the amount of event multiplexing', `ProfileEvent_PerfDataTLBReferences` UInt64 COMMENT 'Data TLB references', `ProfileEvent_PerfDataTLBMisses` UInt64 COMMENT 'Data TLB misses', `ProfileEvent_PerfInstructionTLBReferences` UInt64 COMMENT 'Instruction TLB references', `ProfileEvent_PerfInstructionTLBMisses` UInt64 COMMENT 'Instruction TLB misses', `ProfileEvent_PerfLocalMemoryReferences` UInt64 COMMENT 'Local NUMA node memory reads', `ProfileEvent_PerfLocalMemoryMisses` UInt64 COMMENT 'Local NUMA node memory read misses', `ProfileEvent_CannotWriteToWriteBufferDiscard` UInt64 COMMENT 'Number of stack traces dropped by query profiler or signal handler because pipe is full or cannot write to pipe.', `ProfileEvent_QueryProfilerSignalOverruns` UInt64 COMMENT 'Number of times we drop processing of a query profiler signal due to overrun plus the number of signals that OS has not delivered due to overrun.', `ProfileEvent_QueryProfilerConcurrencyOverruns` UInt64 COMMENT 'Number of times we drop processing of a query profiler signal due to too many concurrent query profilers in other threads, which may indicate overload.', `ProfileEvent_QueryProfilerRuns` UInt64 COMMENT 'Number of times QueryProfiler had been run.', `ProfileEvent_QueryProfilerErrors` UInt64 COMMENT 'Invalid memory accesses during asynchronous stack unwinding.', `ProfileEvent_CreatedLogEntryForMerge` UInt64 COMMENT 'Successfully created log entry to merge parts in ReplicatedMergeTree.', `ProfileEvent_NotCreatedLogEntryForMerge` UInt64 COMMENT 'Log entry to merge parts in ReplicatedMergeTree is not created due to concurrent log update by another replica.', `ProfileEvent_CreatedLogEntryForMutation` UInt64 COMMENT 'Successfully created log entry to mutate parts in ReplicatedMergeTree.', `ProfileEvent_NotCreatedLogEntryForMutation` UInt64 COMMENT 'Log entry to mutate parts in ReplicatedMergeTree is not created due to concurrent log update by another replica.', `ProfileEvent_S3ReadMicroseconds` UInt64 COMMENT 'Time of GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsCount` UInt64 COMMENT 'Number of GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in GET and HEAD requests to S3 storage.', `ProfileEvent_S3ReadRequestsRedirects` UInt64 COMMENT 'Number of redirects in GET and HEAD requests to S3 storage.', `ProfileEvent_S3WriteMicroseconds` UInt64 COMMENT 'Time of POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsCount` UInt64 COMMENT 'Number of POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_S3WriteRequestsRedirects` UInt64 COMMENT 'Number of redirects in POST, DELETE, PUT and PATCH requests to S3 storage.', `ProfileEvent_DiskS3ReadMicroseconds` UInt64 COMMENT 'Time of GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsCount` UInt64 COMMENT 'Number of GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3ReadRequestsRedirects` UInt64 COMMENT 'Number of redirects in GET and HEAD requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteMicroseconds` UInt64 COMMENT 'Time of POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsCount` UInt64 COMMENT 'Number of POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsErrors` UInt64 COMMENT 'Number of non-throttling errors in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsThrottling` UInt64 COMMENT 'Number of 429 and 503 errors in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_DiskS3WriteRequestsRedirects` UInt64 COMMENT 'Number of redirects in POST, DELETE, PUT and PATCH requests to DiskS3 storage.', `ProfileEvent_S3DeleteObjects` UInt64 COMMENT 'Number of S3 API DeleteObject(s) calls.', `ProfileEvent_S3CopyObject` UInt64 COMMENT 'Number of S3 API CopyObject calls.', `ProfileEvent_S3ListObjects` UInt64 COMMENT 'Number of S3 API ListObjects calls.', `ProfileEvent_S3HeadObject` UInt64 COMMENT 'Number of S3 API HeadObject calls.', `ProfileEvent_S3GetObjectAttributes` UInt64 COMMENT 'Number of S3 API GetObjectAttributes calls.', `ProfileEvent_S3CreateMultipartUpload` UInt64 COMMENT 'Number of S3 API CreateMultipartUpload calls.', `ProfileEvent_S3UploadPartCopy` UInt64 COMMENT 'Number of S3 API UploadPartCopy calls.', `ProfileEvent_S3UploadPart` UInt64 COMMENT 'Number of S3 API UploadPart calls.', `ProfileEvent_S3AbortMultipartUpload` UInt64 COMMENT 'Number of S3 API AbortMultipartUpload calls.', `ProfileEvent_S3CompleteMultipartUpload` UInt64 COMMENT 'Number of S3 API CompleteMultipartUpload calls.', `ProfileEvent_S3PutObject` UInt64 COMMENT 'Number of S3 API PutObject calls.', `ProfileEvent_S3GetObject` UInt64 COMMENT 'Number of S3 API GetObject calls.', `ProfileEvent_DiskS3DeleteObjects` UInt64 COMMENT 'Number of DiskS3 API DeleteObject(s) calls.', `ProfileEvent_DiskS3CopyObject` UInt64 COMMENT 'Number of DiskS3 API CopyObject calls.', `ProfileEvent_DiskS3ListObjects` UInt64 COMMENT 'Number of DiskS3 API ListObjects calls.', `ProfileEvent_DiskS3HeadObject` UInt64 COMMENT 'Number of DiskS3 API HeadObject calls.', `ProfileEvent_DiskS3GetObjectAttributes` UInt64 COMMENT 'Number of DiskS3 API GetObjectAttributes calls.', `ProfileEvent_DiskS3CreateMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API CreateMultipartUpload calls.', `ProfileEvent_DiskS3UploadPartCopy` UInt64 COMMENT 'Number of DiskS3 API UploadPartCopy calls.', `ProfileEvent_DiskS3UploadPart` UInt64 COMMENT 'Number of DiskS3 API UploadPart calls.', `ProfileEvent_DiskS3AbortMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API AbortMultipartUpload calls.', `ProfileEvent_DiskS3CompleteMultipartUpload` UInt64 COMMENT 'Number of DiskS3 API CompleteMultipartUpload calls.', `ProfileEvent_DiskS3PutObject` UInt64 COMMENT 'Number of DiskS3 API PutObject calls.', `ProfileEvent_DiskS3GetObject` UInt64 COMMENT 'Number of DiskS3 API GetObject calls.', `ProfileEvent_DiskPlainRewritableAzureDirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for AzureObjectStorage.', `ProfileEvent_DiskPlainRewritableAzureDirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for AzureObjectStorage.', `ProfileEvent_DiskPlainRewritableLocalDirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for LocalObjectStorage.', `ProfileEvent_DiskPlainRewritableLocalDirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for LocalObjectStorage.', `ProfileEvent_DiskPlainRewritableS3DirectoryCreated` UInt64 COMMENT 'Number of directories created by the \'plain_rewritable\' metadata storage for S3ObjectStorage.', `ProfileEvent_DiskPlainRewritableS3DirectoryRemoved` UInt64 COMMENT 'Number of directories removed by the \'plain_rewritable\' metadata storage for S3ObjectStorage.', `ProfileEvent_S3Clients` UInt64 COMMENT 'Number of created S3 clients.', `ProfileEvent_TinyS3Clients` UInt64 COMMENT 'Number of S3 clients copies which reuse an existing auth provider from another client.', `ProfileEvent_EngineFileLikeReadFiles` UInt64 COMMENT 'Number of files read in table engines working with files (like File/S3/URL/HDFS).', `ProfileEvent_ReadBufferFromS3Microseconds` UInt64 COMMENT 'Time spent on reading from S3.', `ProfileEvent_ReadBufferFromS3InitMicroseconds` UInt64 COMMENT 'Time spent initializing connection to S3.', `ProfileEvent_ReadBufferFromS3Bytes` UInt64 COMMENT 'Bytes read from S3.', `ProfileEvent_ReadBufferFromS3RequestsErrors` UInt64 COMMENT 'Number of exceptions while reading from S3.', `ProfileEvent_WriteBufferFromS3Microseconds` UInt64 COMMENT 'Time spent on writing to S3.', `ProfileEvent_WriteBufferFromS3Bytes` UInt64 COMMENT 'Bytes written to S3.', `ProfileEvent_WriteBufferFromS3RequestsErrors` UInt64 COMMENT 'Number of exceptions while writing to S3.', `ProfileEvent_WriteBufferFromS3WaitInflightLimitMicroseconds` UInt64 COMMENT 'Time spent on waiting while some of the current requests are done when its number reached the limit defined by s3_max_inflight_parts_for_one_file.', `ProfileEvent_QueryMemoryLimitExceeded` UInt64 COMMENT 'Number of times when memory limit exceeded for query.', `ProfileEvent_AzureGetObject` UInt64 COMMENT 'Number of Azure API GetObject calls.', `ProfileEvent_AzureUpload` UInt64 COMMENT 'Number of Azure blob storage API Upload calls', `ProfileEvent_AzureStageBlock` UInt64 COMMENT 'Number of Azure blob storage API StageBlock calls', `ProfileEvent_AzureCommitBlockList` UInt64 COMMENT 'Number of Azure blob storage API CommitBlockList calls', `ProfileEvent_AzureCopyObject` UInt64 COMMENT 'Number of Azure blob storage API CopyObject calls', `ProfileEvent_AzureDeleteObjects` UInt64 COMMENT 'Number of Azure blob storage API DeleteObject(s) calls.', `ProfileEvent_AzureListObjects` UInt64 COMMENT 'Number of Azure blob storage API ListObjects calls.', `ProfileEvent_AzureGetProperties` UInt64 COMMENT 'Number of Azure blob storage API GetProperties calls.', `ProfileEvent_AzureCreateContainer` UInt64 COMMENT 'Number of Azure blob storage API CreateContainer calls.', `ProfileEvent_DiskAzureGetObject` UInt64 COMMENT 'Number of Disk Azure API GetObject calls.', `ProfileEvent_DiskAzureUpload` UInt64 COMMENT 'Number of Disk Azure blob storage API Upload calls', `ProfileEvent_DiskAzureStageBlock` UInt64 COMMENT 'Number of Disk Azure blob storage API StageBlock calls', `ProfileEvent_DiskAzureCommitBlockList` UInt64 COMMENT 'Number of Disk Azure blob storage API CommitBlockList calls', `ProfileEvent_DiskAzureCopyObject` UInt64 COMMENT 'Number of Disk Azure blob storage API CopyObject calls', `ProfileEvent_DiskAzureListObjects` UInt64 COMMENT 'Number of Disk Azure blob storage API ListObjects calls.', `ProfileEvent_DiskAzureDeleteObjects` UInt64 COMMENT 'Number of Azure blob storage API DeleteObject(s) calls.', `ProfileEvent_DiskAzureGetProperties` UInt64 COMMENT 'Number of Disk Azure blob storage API GetProperties calls.', `ProfileEvent_DiskAzureCreateContainer` UInt64 COMMENT 'Number of Disk Azure blob storage API CreateContainer calls.', `ProfileEvent_ReadBufferFromAzureMicroseconds` UInt64 COMMENT 'Time spent on reading from Azure.', `ProfileEvent_ReadBufferFromAzureInitMicroseconds` UInt64 COMMENT 'Time spent initializing connection to Azure.', `ProfileEvent_ReadBufferFromAzureBytes` UInt64 COMMENT 'Bytes read from Azure.', `ProfileEvent_ReadBufferFromAzureRequestsErrors` UInt64 COMMENT 'Number of exceptions while reading from Azure', `ProfileEvent_CachedReadBufferReadFromCacheHits` UInt64 COMMENT 'Number of times the read from filesystem cache hit the cache.', `ProfileEvent_CachedReadBufferReadFromCacheMisses` UInt64 COMMENT 'Number of times the read from filesystem cache miss the cache.', `ProfileEvent_CachedReadBufferReadFromSourceMicroseconds` UInt64 COMMENT 'Time reading from filesystem cache source (from remote filesystem, etc)', `ProfileEvent_CachedReadBufferReadFromCacheMicroseconds` UInt64 COMMENT 'Time reading from filesystem cache', `ProfileEvent_CachedReadBufferReadFromSourceBytes` UInt64 COMMENT 'Bytes read from filesystem cache source (from remote fs, etc)', `ProfileEvent_CachedReadBufferReadFromCacheBytes` UInt64 COMMENT 'Bytes read from filesystem cache', `ProfileEvent_CachedReadBufferPredownloadedBytes` UInt64 COMMENT 'Bytes read from filesystem cache source. Cache segments are read from left to right as a whole, it might be that we need to predownload some part of the segment irrelevant for the current task just to get to the needed data', `ProfileEvent_CachedReadBufferCacheWriteBytes` UInt64 COMMENT 'Bytes written from source (remote fs, etc) to filesystem cache', `ProfileEvent_CachedReadBufferCacheWriteMicroseconds` UInt64 COMMENT 'Time spent writing data into filesystem cache', `ProfileEvent_CachedReadBufferCreateBufferMicroseconds` UInt64 COMMENT 'Prepare buffer time', `ProfileEvent_CachedWriteBufferCacheWriteBytes` UInt64 COMMENT 'Bytes written from source (remote fs, etc) to filesystem cache', `ProfileEvent_CachedWriteBufferCacheWriteMicroseconds` UInt64 COMMENT 'Time spent writing data into filesystem cache', `ProfileEvent_FilesystemCacheLoadMetadataMicroseconds` UInt64 COMMENT 'Time spent loading filesystem cache metadata', `ProfileEvent_FilesystemCacheEvictedBytes` UInt64 COMMENT 'Number of bytes evicted from filesystem cache', `ProfileEvent_FilesystemCacheEvictedFileSegments` UInt64 COMMENT 'Number of file segments evicted from filesystem cache', `ProfileEvent_FilesystemCacheBackgroundDownloadQueuePush` UInt64 COMMENT 'Number of file segments sent for background download in filesystem cache', `ProfileEvent_FilesystemCacheEvictionSkippedFileSegments` UInt64 COMMENT 'Number of file segments skipped for eviction because of being in unreleasable state', `ProfileEvent_FilesystemCacheEvictionSkippedEvictingFileSegments` UInt64 COMMENT 'Number of file segments skipped for eviction because of being in evicting state', `ProfileEvent_FilesystemCacheEvictionTries` UInt64 COMMENT 'Number of filesystem cache eviction attempts', `ProfileEvent_FilesystemCacheLockKeyMicroseconds` UInt64 COMMENT 'Lock cache key time', `ProfileEvent_FilesystemCacheLockMetadataMicroseconds` UInt64 COMMENT 'Lock filesystem cache metadata time', `ProfileEvent_FilesystemCacheLockCacheMicroseconds` UInt64 COMMENT 'Lock filesystem cache time', `ProfileEvent_FilesystemCacheReserveMicroseconds` UInt64 COMMENT 'Filesystem cache space reservation time', `ProfileEvent_FilesystemCacheEvictMicroseconds` UInt64 COMMENT 'Filesystem cache eviction time', `ProfileEvent_FilesystemCacheGetOrSetMicroseconds` UInt64 COMMENT 'Filesystem cache getOrSet() time', `ProfileEvent_FilesystemCacheGetMicroseconds` UInt64 COMMENT 'Filesystem cache get() time', `ProfileEvent_FileSegmentWaitMicroseconds` UInt64 COMMENT 'Wait on DOWNLOADING state', `ProfileEvent_FileSegmentCompleteMicroseconds` UInt64 COMMENT 'Duration of FileSegment::complete() in filesystem cache', `ProfileEvent_FileSegmentLockMicroseconds` UInt64 COMMENT 'Lock file segment time', `ProfileEvent_FileSegmentWriteMicroseconds` UInt64 COMMENT 'File segment write() time', `ProfileEvent_FileSegmentUseMicroseconds` UInt64 COMMENT 'File segment use() time', `ProfileEvent_FileSegmentRemoveMicroseconds` UInt64 COMMENT 'File segment remove() time', `ProfileEvent_FileSegmentHolderCompleteMicroseconds` UInt64 COMMENT 'File segments holder complete() time', `ProfileEvent_FileSegmentFailToIncreasePriority` UInt64 COMMENT 'Number of times the priority was not increased due to a high contention on the cache lock', `ProfileEvent_FilesystemCacheFailToReserveSpaceBecauseOfLockContention` UInt64 COMMENT 'Number of times space reservation was skipped due to a high contention on the cache lock', `ProfileEvent_FilesystemCacheFailToReserveSpaceBecauseOfCacheResize` UInt64 COMMENT 'Number of times space reservation was skipped due to the cache is being resized', `ProfileEvent_FilesystemCacheHoldFileSegments` UInt64 COMMENT 'Filesystem cache file segments count, which were hold', `ProfileEvent_FilesystemCacheUnusedHoldFileSegments` UInt64 COMMENT 'Filesystem cache file segments count, which were hold, but not used (because of seek or LIMIT n, etc)', `ProfileEvent_FilesystemCacheFreeSpaceKeepingThreadRun` UInt64 COMMENT 'Number of times background thread executed free space keeping job', `ProfileEvent_FilesystemCacheFreeSpaceKeepingThreadWorkMilliseconds` UInt64 COMMENT 'Time for which background thread executed free space keeping job', `ProfileEvent_RemoteFSSeeks` UInt64 COMMENT 'Total number of seeks for async buffer', `ProfileEvent_RemoteFSPrefetches` UInt64 COMMENT 'Number of prefetches made with asynchronous reading from remote filesystem', `ProfileEvent_RemoteFSCancelledPrefetches` UInt64 COMMENT 'Number of cancelled prefecthes (because of seek)', `ProfileEvent_RemoteFSUnusedPrefetches` UInt64 COMMENT 'Number of prefetches pending at buffer destruction', `ProfileEvent_RemoteFSPrefetchedReads` UInt64 COMMENT 'Number of reads from prefecthed buffer', `ProfileEvent_RemoteFSPrefetchedBytes` UInt64 COMMENT 'Number of bytes from prefecthed buffer', `ProfileEvent_RemoteFSUnprefetchedReads` UInt64 COMMENT 'Number of reads from unprefetched buffer', `ProfileEvent_RemoteFSUnprefetchedBytes` UInt64 COMMENT 'Number of bytes from unprefetched buffer', `ProfileEvent_RemoteFSLazySeeks` UInt64 COMMENT 'Number of lazy seeks', `ProfileEvent_RemoteFSSeeksWithReset` UInt64 COMMENT 'Number of seeks which lead to a new connection', `ProfileEvent_RemoteFSBuffers` UInt64 COMMENT 'Number of buffers created for asynchronous reading from remote filesystem', `ProfileEvent_MergeTreePrefetchedReadPoolInit` UInt64 COMMENT 'Time spent preparing tasks in MergeTreePrefetchedReadPool', `ProfileEvent_WaitPrefetchTaskMicroseconds` UInt64 COMMENT 'Time spend waiting for prefetched reader', `ProfileEvent_ThreadpoolReaderTaskMicroseconds` UInt64 COMMENT 'Time spent getting the data in asynchronous reading', `ProfileEvent_ThreadpoolReaderPrepareMicroseconds` UInt64 COMMENT 'Time spent on preparation (e.g. call to reader seek() method)', `ProfileEvent_ThreadpoolReaderReadBytes` UInt64 COMMENT 'Bytes read from a threadpool task in asynchronous reading', `ProfileEvent_ThreadpoolReaderSubmit` UInt64 COMMENT 'Bytes read from a threadpool task in asynchronous reading', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronously` UInt64 COMMENT 'How many times we haven\'t scheduled a task on the thread pool and read synchronously instead', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronouslyBytes` UInt64 COMMENT 'How many bytes were read synchronously', `ProfileEvent_ThreadpoolReaderSubmitReadSynchronouslyMicroseconds` UInt64 COMMENT 'How much time we spent reading synchronously', `ProfileEvent_ThreadpoolReaderSubmitLookupInCacheMicroseconds` UInt64 COMMENT 'How much time we spent checking if content is cached', `ProfileEvent_AsynchronousReaderIgnoredBytes` UInt64 COMMENT 'Number of bytes ignored during asynchronous reading', `ProfileEvent_FileSegmentWaitReadBufferMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend waiting for internal read buffer (includes cache waiting)', `ProfileEvent_FileSegmentReadMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend reading from file', `ProfileEvent_FileSegmentCacheWriteMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spend writing data to cache', `ProfileEvent_FileSegmentPredownloadMicroseconds` UInt64 COMMENT 'Metric per file segment. Time spent pre-downloading data to cache (pre-downloading - finishing file segment download (after someone who failed to do that) up to the point current thread was requested to do)', `ProfileEvent_FileSegmentUsedBytes` UInt64 COMMENT 'Metric per file segment. How many bytes were actually used from current file segment', `ProfileEvent_ReadBufferSeekCancelConnection` UInt64 COMMENT 'Number of seeks which lead to new connection (s3, http)', `ProfileEvent_SleepFunctionCalls` UInt64 COMMENT 'Number of times a sleep function (sleep, sleepEachRow) has been called.', `ProfileEvent_SleepFunctionMicroseconds` UInt64 COMMENT 'Time set to sleep in a sleep function (sleep, sleepEachRow).', `ProfileEvent_SleepFunctionElapsedMicroseconds` UInt64 COMMENT 'Time spent sleeping in a sleep function (sleep, sleepEachRow).', `ProfileEvent_ThreadPoolReaderPageCacheHit` UInt64 COMMENT 'Number of times the read inside ThreadPoolReader was done from the page cache.', `ProfileEvent_ThreadPoolReaderPageCacheHitBytes` UInt64 COMMENT 'Number of bytes read inside ThreadPoolReader when it was done from the page cache.', `ProfileEvent_ThreadPoolReaderPageCacheHitElapsedMicroseconds` UInt64 COMMENT 'Time spent reading data from page cache in ThreadPoolReader.', `ProfileEvent_ThreadPoolReaderPageCacheMiss` UInt64 COMMENT 'Number of times the read inside ThreadPoolReader was not done from page cache and was hand off to thread pool.', `ProfileEvent_ThreadPoolReaderPageCacheMissBytes` UInt64 COMMENT 'Number of bytes read inside ThreadPoolReader when read was not done from page cache and was hand off to thread pool.', `ProfileEvent_ThreadPoolReaderPageCacheMissElapsedMicroseconds` UInt64 COMMENT 'Time spent reading data inside the asynchronous job in ThreadPoolReader - when read was not done from the page cache.', `ProfileEvent_AsynchronousReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for asynchronous reads in asynchronous local read.', `ProfileEvent_SynchronousReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for synchronous reads in asynchronous local read.', `ProfileEvent_AsynchronousRemoteReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for asynchronous remote reads.', `ProfileEvent_SynchronousRemoteReadWaitMicroseconds` UInt64 COMMENT 'Time spent in waiting for synchronous remote reads.', `ProfileEvent_ExternalDataSourceLocalCacheReadBytes` UInt64 COMMENT 'Bytes read from local cache buffer in RemoteReadBufferCache', `ProfileEvent_MainConfigLoads` UInt64 COMMENT 'Number of times the main configuration was reloaded.', `ProfileEvent_AggregationPreallocatedElementsInHashTables` UInt64 COMMENT 'How many elements were preallocated in hash tables for aggregation.', `ProfileEvent_AggregationHashTablesInitializedAsTwoLevel` UInt64 COMMENT 'How many hash tables were inited as two-level for aggregation.', `ProfileEvent_AggregationOptimizedEqualRangesOfKeys` UInt64 COMMENT 'For how many blocks optimization of equal ranges of keys was applied', `ProfileEvent_HashJoinPreallocatedElementsInHashTables` UInt64 COMMENT 'How many elements were preallocated in hash tables for hash join.', `ProfileEvent_MetadataFromKeeperCacheHit` UInt64 COMMENT 'Number of times an object storage metadata request was answered from cache without making request to Keeper', `ProfileEvent_MetadataFromKeeperCacheMiss` UInt64 COMMENT 'Number of times an object storage metadata request had to be answered from Keeper', `ProfileEvent_MetadataFromKeeperCacheUpdateMicroseconds` UInt64 COMMENT 'Total time spent in updating the cache including waiting for responses from Keeper', `ProfileEvent_MetadataFromKeeperUpdateCacheOneLevel` UInt64 COMMENT 'Number of times a cache update for one level of directory tree was done', `ProfileEvent_MetadataFromKeeperTransactionCommit` UInt64 COMMENT 'Number of times metadata transaction commit was attempted', `ProfileEvent_MetadataFromKeeperTransactionCommitRetry` UInt64 COMMENT 'Number of times metadata transaction commit was retried', `ProfileEvent_MetadataFromKeeperCleanupTransactionCommit` UInt64 COMMENT 'Number of times metadata transaction commit for deleted objects cleanup was attempted', `ProfileEvent_MetadataFromKeeperCleanupTransactionCommitRetry` UInt64 COMMENT 'Number of times metadata transaction commit for deleted objects cleanup was retried', `ProfileEvent_MetadataFromKeeperOperations` UInt64 COMMENT 'Number of times a request was made to Keeper', `ProfileEvent_MetadataFromKeeperIndividualOperations` UInt64 COMMENT 'Number of paths read or written by single or multi requests to Keeper', `ProfileEvent_MetadataFromKeeperReconnects` UInt64 COMMENT 'Number of times a reconnect to Keeper was done', `ProfileEvent_MetadataFromKeeperBackgroundCleanupObjects` UInt64 COMMENT 'Number of times a old deleted object clean up was performed by background task', `ProfileEvent_MetadataFromKeeperBackgroundCleanupTransactions` UInt64 COMMENT 'Number of times old transaction idempotency token was cleaned up by background task', `ProfileEvent_MetadataFromKeeperBackgroundCleanupErrors` UInt64 COMMENT 'Number of times an error was encountered in background cleanup task', `ProfileEvent_SharedMergeTreeMetadataCacheHintLoadedFromCache` UInt64 COMMENT 'Number of times metadata cache hint was found without going to Keeper', `ProfileEvent_KafkaRebalanceRevocations` UInt64 COMMENT 'Number of partition revocations (the first stage of consumer group rebalance)', `ProfileEvent_KafkaRebalanceAssignments` UInt64 COMMENT 'Number of partition assignments (the final stage of consumer group rebalance)', `ProfileEvent_KafkaRebalanceErrors` UInt64 COMMENT 'Number of failed consumer group rebalances', `ProfileEvent_KafkaMessagesPolled` UInt64 COMMENT 'Number of Kafka messages polled from librdkafka to ClickHouse', `ProfileEvent_KafkaMessagesRead` UInt64 COMMENT 'Number of Kafka messages already processed by ClickHouse', `ProfileEvent_KafkaMessagesFailed` UInt64 COMMENT 'Number of Kafka messages ClickHouse failed to parse', `ProfileEvent_KafkaRowsRead` UInt64 COMMENT 'Number of rows parsed from Kafka messages', `ProfileEvent_KafkaRowsRejected` UInt64 COMMENT 'Number of parsed rows which were later rejected (due to rebalances / errors or similar reasons). Those rows will be consumed again after the rebalance.', `ProfileEvent_KafkaDirectReads` UInt64 COMMENT 'Number of direct selects from Kafka tables since server start', `ProfileEvent_KafkaBackgroundReads` UInt64 COMMENT 'Number of background reads populating materialized views from Kafka since server start', `ProfileEvent_KafkaCommits` UInt64 COMMENT 'Number of successful commits of consumed offsets to Kafka (normally should be the same as KafkaBackgroundReads)', `ProfileEvent_KafkaCommitFailures` UInt64 COMMENT 'Number of failed commits of consumed offsets to Kafka (usually is a sign of some data duplication)', `ProfileEvent_KafkaConsumerErrors` UInt64 COMMENT 'Number of errors reported by librdkafka during polls', `ProfileEvent_KafkaWrites` UInt64 COMMENT 'Number of writes (inserts) to Kafka tables ', `ProfileEvent_KafkaRowsWritten` UInt64 COMMENT 'Number of rows inserted into Kafka tables', `ProfileEvent_KafkaProducerFlushes` UInt64 COMMENT 'Number of explicit flushes to Kafka producer', `ProfileEvent_KafkaMessagesProduced` UInt64 COMMENT 'Number of messages produced to Kafka', `ProfileEvent_KafkaProducerErrors` UInt64 COMMENT 'Number of errors during producing the messages to Kafka', `ProfileEvent_ScalarSubqueriesGlobalCacheHit` UInt64 COMMENT 'Number of times a read from a scalar subquery was done using the global cache', `ProfileEvent_ScalarSubqueriesLocalCacheHit` UInt64 COMMENT 'Number of times a read from a scalar subquery was done using the local cache', `ProfileEvent_ScalarSubqueriesCacheMiss` UInt64 COMMENT 'Number of times a read from a scalar subquery was not cached and had to be calculated completely', `ProfileEvent_SchemaInferenceCacheHits` UInt64 COMMENT 'Number of times the requested source is found in schema cache', `ProfileEvent_SchemaInferenceCacheSchemaHits` UInt64 COMMENT 'Number of times the schema is found in schema cache during schema inference', `ProfileEvent_SchemaInferenceCacheNumRowsHits` UInt64 COMMENT 'Number of times the number of rows is found in schema cache during count from files', `ProfileEvent_SchemaInferenceCacheMisses` UInt64 COMMENT 'Number of times the requested source is not in schema cache', `ProfileEvent_SchemaInferenceCacheSchemaMisses` UInt64 COMMENT 'Number of times the requested source is in cache but the schema is not in cache during schema inference', `ProfileEvent_SchemaInferenceCacheNumRowsMisses` UInt64 COMMENT 'Number of times the requested source is in cache but the number of rows is not in cache while count from files', `ProfileEvent_SchemaInferenceCacheEvictions` UInt64 COMMENT 'Number of times a schema from cache was evicted due to overflow', `ProfileEvent_SchemaInferenceCacheInvalidations` UInt64 COMMENT 'Number of times a schema in cache became invalid due to changes in data', `ProfileEvent_KeeperPacketsSent` UInt64 COMMENT 'Packets sent by keeper server', `ProfileEvent_KeeperPacketsReceived` UInt64 COMMENT 'Packets received by keeper server', `ProfileEvent_KeeperRequestTotal` UInt64 COMMENT 'Total requests number on keeper server', `ProfileEvent_KeeperLatency` UInt64 COMMENT 'Keeper latency', `ProfileEvent_KeeperTotalElapsedMicroseconds` UInt64 COMMENT 'Keeper total latency for a single request', `ProfileEvent_KeeperProcessElapsedMicroseconds` UInt64 COMMENT 'Keeper commit latency for a single request', `ProfileEvent_KeeperPreprocessElapsedMicroseconds` UInt64 COMMENT 'Keeper preprocessing latency for a single reuquest', `ProfileEvent_KeeperStorageLockWaitMicroseconds` UInt64 COMMENT 'Time spent waiting for acquiring Keeper storage lock', `ProfileEvent_KeeperCommitWaitElapsedMicroseconds` UInt64 COMMENT 'Time spent waiting for certain log to be committed', `ProfileEvent_KeeperBatchMaxCount` UInt64 COMMENT 'Number of times the size of batch was limited by the amount', `ProfileEvent_KeeperBatchMaxTotalSize` UInt64 COMMENT 'Number of times the size of batch was limited by the total bytes size', `ProfileEvent_KeeperCommits` UInt64 COMMENT 'Number of successful commits', `ProfileEvent_KeeperCommitsFailed` UInt64 COMMENT 'Number of failed commits', `ProfileEvent_KeeperSnapshotCreations` UInt64 COMMENT 'Number of snapshots creations', `ProfileEvent_KeeperSnapshotCreationsFailed` UInt64 COMMENT 'Number of failed snapshot creations', `ProfileEvent_KeeperSnapshotApplys` UInt64 COMMENT 'Number of snapshot applying', `ProfileEvent_KeeperSnapshotApplysFailed` UInt64 COMMENT 'Number of failed snapshot applying', `ProfileEvent_KeeperReadSnapshot` UInt64 COMMENT 'Number of snapshot read(serialization)', `ProfileEvent_KeeperSaveSnapshot` UInt64 COMMENT 'Number of snapshot save', `ProfileEvent_KeeperCreateRequest` UInt64 COMMENT 'Number of create requests', `ProfileEvent_KeeperRemoveRequest` UInt64 COMMENT 'Number of remove requests', `ProfileEvent_KeeperSetRequest` UInt64 COMMENT 'Number of set requests', `ProfileEvent_KeeperReconfigRequest` UInt64 COMMENT 'Number of reconfig requests', `ProfileEvent_KeeperCheckRequest` UInt64 COMMENT 'Number of check requests', `ProfileEvent_KeeperMultiRequest` UInt64 COMMENT 'Number of multi requests', `ProfileEvent_KeeperMultiReadRequest` UInt64 COMMENT 'Number of multi read requests', `ProfileEvent_KeeperGetRequest` UInt64 COMMENT 'Number of get requests', `ProfileEvent_KeeperListRequest` UInt64 COMMENT 'Number of list requests', `ProfileEvent_KeeperExistsRequest` UInt64 COMMENT 'Number of exists requests', `ProfileEvent_OverflowBreak` UInt64 COMMENT 'Number of times, data processing was cancelled by query complexity limitation with setting \'*_overflow_mode\' = \'break\' and the result is incomplete.', `ProfileEvent_OverflowThrow` UInt64 COMMENT 'Number of times, data processing was cancelled by query complexity limitation with setting \'*_overflow_mode\' = \'throw\' and exception was thrown.', `ProfileEvent_OverflowAny` UInt64 COMMENT 'Number of times approximate GROUP BY was in effect: when aggregation was performed only on top of first \'max_rows_to_group_by\' unique keys and other keys were ignored due to \'group_by_overflow_mode\' = \'any\'.', `ProfileEvent_S3QueueSetFileProcessingMicroseconds` UInt64 COMMENT 'Time spent to set file as processing', `ProfileEvent_S3QueueSetFileProcessedMicroseconds` UInt64 COMMENT 'Time spent to set file as processed', `ProfileEvent_S3QueueSetFileFailedMicroseconds` UInt64 COMMENT 'Time spent to set file as failed', `ProfileEvent_ObjectStorageQueueFailedFiles` UInt64 COMMENT 'Number of files which failed to be processed', `ProfileEvent_ObjectStorageQueueProcessedFiles` UInt64 COMMENT 'Number of files which were processed', `ProfileEvent_ObjectStorageQueueCleanupMaxSetSizeOrTTLMicroseconds` UInt64 COMMENT 'Time spent to set file as failed', `ProfileEvent_ObjectStorageQueuePullMicroseconds` UInt64 COMMENT 'Time spent to read file data', `ProfileEvent_ObjectStorageQueueLockLocalFileStatusesMicroseconds` UInt64 COMMENT 'Time spent to lock local file statuses', `ProfileEvent_ObjectStorageQueueFailedToBatchSetProcessing` UInt64 COMMENT 'Number of times batched set processing request failed', `ProfileEvent_ObjectStorageQueueTrySetProcessingRequests` UInt64 COMMENT 'The number of times we tried to make set processing request', `ProfileEvent_ObjectStorageQueueTrySetProcessingSucceeded` UInt64 COMMENT 'The number of times we successfully set file as processing', `ProfileEvent_ObjectStorageQueueTrySetProcessingFailed` UInt64 COMMENT 'The number of times we unsuccessfully set file as processing', `ProfileEvent_ObjectStorageQueueListedFiles` UInt64 COMMENT 'Number of listed files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueFilteredFiles` UInt64 COMMENT 'Number of filtered files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueReadFiles` UInt64 COMMENT 'Number of read files (not equal to the number of actually inserted files)', `ProfileEvent_ObjectStorageQueueReadRows` UInt64 COMMENT 'Number of read rows (not equal to the number of actually inserted rows)', `ProfileEvent_ObjectStorageQueueReadBytes` UInt64 COMMENT 'Number of read bytes (not equal to the number of actually inserted bytes)', `ProfileEvent_ObjectStorageQueueExceptionsDuringRead` UInt64 COMMENT 'Number of exceptions during read in S3(Azure)Queue', `ProfileEvent_ObjectStorageQueueExceptionsDuringInsert` UInt64 COMMENT 'Number of exceptions during insert in S3(Azure)Queue', `ProfileEvent_ObjectStorageQueueRemovedObjects` UInt64 COMMENT 'Number of objects removed as part of after_processing = delete', `ProfileEvent_ObjectStorageQueueInsertIterations` UInt64 COMMENT 'Number of insert iterations', `ProfileEvent_ObjectStorageQueueCommitRequests` UInt64 COMMENT 'Number of keeper requests to commit files as either failed or processed', `ProfileEvent_ObjectStorageQueueSuccessfulCommits` UInt64 COMMENT 'Number of successful keeper commits', `ProfileEvent_ObjectStorageQueueUnsuccessfulCommits` UInt64 COMMENT 'Number of unsuccessful keeper commits', `ProfileEvent_ObjectStorageQueueCancelledFiles` UInt64 COMMENT 'Number cancelled files in StorageS3(Azure)Queue', `ProfileEvent_ObjectStorageQueueProcessedRows` UInt64 COMMENT 'Number of processed rows in StorageS3(Azure)Queue', `ProfileEvent_ServerStartupMilliseconds` UInt64 COMMENT 'Time elapsed from starting server to listening to sockets in milliseconds', `ProfileEvent_IOUringSQEsSubmitted` UInt64 COMMENT 'Total number of io_uring SQEs submitted', `ProfileEvent_IOUringSQEsResubmitsAsync` UInt64 COMMENT 'Total number of asynchronous io_uring SQE resubmits performed', `ProfileEvent_IOUringSQEsResubmitsSync` UInt64 COMMENT 'Total number of synchronous io_uring SQE resubmits performed', `ProfileEvent_IOUringCQEsCompleted` UInt64 COMMENT 'Total number of successfully completed io_uring CQEs', `ProfileEvent_IOUringCQEsFailed` UInt64 COMMENT 'Total number of completed io_uring CQEs with failures', `ProfileEvent_BackupsOpenedForRead` UInt64 COMMENT 'Number of backups opened for reading', `ProfileEvent_BackupsOpenedForWrite` UInt64 COMMENT 'Number of backups opened for writing', `ProfileEvent_BackupsOpenedForUnlock` UInt64 COMMENT 'Number of backups opened for unlocking', `ProfileEvent_BackupReadMetadataMicroseconds` UInt64 COMMENT 'Time spent reading backup metadata from .backup file', `ProfileEvent_BackupWriteMetadataMicroseconds` UInt64 COMMENT 'Time spent writing backup metadata to .backup file', `ProfileEvent_BackupEntriesCollectorMicroseconds` UInt64 COMMENT 'Time spent making backup entries', `ProfileEvent_BackupEntriesCollectorForTablesDataMicroseconds` UInt64 COMMENT 'Time spent making backup entries for tables data', `ProfileEvent_BackupEntriesCollectorRunPostTasksMicroseconds` UInt64 COMMENT 'Time spent running post tasks after making backup entries', `ProfileEvent_BackupPreparingFileInfosMicroseconds` UInt64 COMMENT 'Time spent preparing file infos for backup entries', `ProfileEvent_BackupReadLocalFilesToCalculateChecksums` UInt64 COMMENT 'Number of files read locally to calculate checksums for backup entries', `ProfileEvent_BackupReadLocalBytesToCalculateChecksums` UInt64 COMMENT 'Total size of files read locally to calculate checksums for backup entries', `ProfileEvent_BackupReadRemoteFilesToCalculateChecksums` UInt64 COMMENT 'Number of files read from remote disks to calculate checksums for backup entries', `ProfileEvent_BackupReadRemoteBytesToCalculateChecksums` UInt64 COMMENT 'Total size of files read from remote disks to calculate checksums for backup entries', `ProfileEvent_BackupLockFileReads` UInt64 COMMENT 'How many times the \'.lock\' file was read while making backup', `ProfileEvent_RestorePartsSkippedFiles` UInt64 COMMENT 'Number of files skipped while restoring parts', `ProfileEvent_RestorePartsSkippedBytes` UInt64 COMMENT 'Total size of files skipped while restoring parts', `ProfileEvent_ReadTaskRequestsReceived` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the initiator server side.', `ProfileEvent_MergeTreeReadTaskRequestsReceived` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the initiator server side.', `ProfileEvent_ReadTaskRequestsSent` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `ProfileEvent_MergeTreeReadTaskRequestsSent` UInt64 COMMENT 'The number of callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergeTreeAllRangesAnnouncementsSent` UInt64 COMMENT 'The number of announcements sent from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_ReadTaskRequestsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in callbacks requested from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `ProfileEvent_MergeTreeReadTaskRequestsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in callbacks requested from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergeTreeAllRangesAnnouncementsSentElapsedMicroseconds` UInt64 COMMENT 'Time spent in sending the announcement from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `ProfileEvent_MergerMutatorsGetPartsForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to take data parts snapshot to build ranges from them.', `ProfileEvent_MergerMutatorPrepareRangesForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to prepare parts ranges which can be merged according to merge predicate.', `ProfileEvent_MergerMutatorSelectPartsForMergeElapsedMicroseconds` UInt64 COMMENT 'Time spent to select parts from ranges which can be merged.', `ProfileEvent_MergerMutatorRangesForMergeCount` UInt64 COMMENT 'Amount of candidate ranges for merge', `ProfileEvent_MergerMutatorPartsInRangesForMergeCount` UInt64 COMMENT 'Amount of candidate parts for merge', `ProfileEvent_MergerMutatorSelectRangePartsCount` UInt64 COMMENT 'Amount of parts in selected range for merge', `ProfileEvent_ConnectionPoolIsFullMicroseconds` UInt64 COMMENT 'Total time spent waiting for a slot in connection pool.', `ProfileEvent_AsyncLoaderWaitMicroseconds` UInt64 COMMENT 'Total time a query was waiting for async loader jobs.', `ProfileEvent_DistrCacheServerSwitches` UInt64 COMMENT 'Distributed Cache read buffer event. Number of server switches between distributed cache servers in read/write-through cache', `ProfileEvent_DistrCacheReadMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent reading from distributed cache', `ProfileEvent_DistrCacheFallbackReadMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spend reading from fallback buffer instead of distributed cache', `ProfileEvent_DistrCachePrecomputeRangesMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent to precompute read ranges', `ProfileEvent_DistrCacheNextImplMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spend in ReadBufferFromDistributedCache::nextImpl', `ProfileEvent_DistrCacheStartRangeMicroseconds` UInt64 COMMENT 'Distributed Cache read buffer event. Time spent to start a new read range with distributed cache', `ProfileEvent_DistrCacheIgnoredBytesWhileWaitingProfileEvents` UInt64 COMMENT 'Distributed Cache read buffer event. Ignored bytes while waiting for profile events in distributed cache', `ProfileEvent_DistrCacheRangeChange` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we changed read range because of seek/last_position change', `ProfileEvent_DistrCacheGetResponseMicroseconds` UInt64 COMMENT 'Distributed Cache client event. Time spend to wait for response from distributed cache', `ProfileEvent_DistrCacheReadErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors during read', `ProfileEvent_DistrCacheMakeRequestErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors when making a request', `ProfileEvent_DistrCacheReceiveResponseErrors` UInt64 COMMENT 'Distributed Cache client event. Number of distributed cache errors when receiving response a request', `ProfileEvent_DistrCachePackets` UInt64 COMMENT 'Distributed Cache client event. Total number of packets received from distributed cache', `ProfileEvent_DistrCachePacketsBytes` UInt64 COMMENT 'Distributed Cache client event. The number of bytes in Data packets which were not ignored', `ProfileEvent_DistrCacheUnusedPackets` UInt64 COMMENT 'Distributed Cache client event. Number of skipped unused packets from distributed cache', `ProfileEvent_DistrCacheUnusedPacketsBytes` UInt64 COMMENT 'Distributed Cache client event. The number of bytes in Data packets which were ignored', `ProfileEvent_DistrCacheUnusedPacketsBufferAllocations` UInt64 COMMENT 'Distributed Cache client event. The number of extra buffer allocations in case we could not reuse existing buffer', `ProfileEvent_DistrCacheLockRegistryMicroseconds` UInt64 COMMENT 'Distributed Cache registry event. Time spent to take DistributedCacheRegistry lock', `ProfileEvent_DistrCacheRegistryUpdateMicroseconds` UInt64 COMMENT 'Distributed Cache registry event. Time spent updating distributed cache registry', `ProfileEvent_DistrCacheRegistryUpdates` UInt64 COMMENT 'Distributed Cache registry event. Number of distributed cache registry updates', `ProfileEvent_DistrCacheHashRingRebuilds` UInt64 COMMENT 'Distributed Cache registry event. Number of distributed cache hash ring rebuilds', `ProfileEvent_DistrCacheReadBytesFromCache` UInt64 COMMENT 'Distributed Cache read buffer event. Bytes read from distributed cache', `ProfileEvent_DistrCacheReadBytesFromFallbackBuffer` UInt64 COMMENT 'Distributed Cache read buffer event. Bytes read from fallback buffer', `ProfileEvent_DistrCacheRangeResetBackward` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we reset read range because of seek/last_position change', `ProfileEvent_DistrCacheRangeResetForward` UInt64 COMMENT 'Distributed Cache read buffer event. Number of times we reset read range because of seek/last_position change', `ProfileEvent_DistrCacheOpenedConnections` UInt64 COMMENT 'Distributed Cache connection event. The number of open connections to distributed cache', `ProfileEvent_DistrCacheReusedConnections` UInt64 COMMENT 'Distributed Cache connection event. The number of reused connections to distributed cache', `ProfileEvent_DistrCacheOpenedConnectionsBypassingPool` UInt64 COMMENT 'Distributed Cache connection event. The number of open connections to distributed cache bypassing pool', `ProfileEvent_DistrCacheConnectMicroseconds` UInt64 COMMENT 'Distributed Cache connection event. The time spent to connect to distributed cache', `ProfileEvent_DistrCacheConnectAttempts` UInt64 COMMENT 'Distributed Cache connection event. The number of connection attempts to distributed cache', `ProfileEvent_DistrCacheGetClientMicroseconds` UInt64 COMMENT 'Distributed Cache connection event. Time spent getting client for distributed cache', `ProfileEvent_DistrCacheServerProcessRequestMicroseconds` UInt64 COMMENT 'Distributed Cache server event. Time spent processing request on DistributedCache server side', `ProfileEvent_DistrCacheServerStartRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of StartRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerContinueRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of ContinueRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerEndRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of EndRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerAckRequestPackets` UInt64 COMMENT 'Distributed Cache server event. Number of AckRequest packets in DistributedCacheServer', `ProfileEvent_DistrCacheServerNewS3CachedClients` UInt64 COMMENT 'Distributed Cache server event. The number of new cached s3 clients', `ProfileEvent_DistrCacheServerReusedS3CachedClients` UInt64 COMMENT 'Distributed Cache server event. The number of reused cached s3 clients', `ProfileEvent_LogTest` UInt64 COMMENT 'Number of log messages with level Test', `ProfileEvent_LogTrace` UInt64 COMMENT 'Number of log messages with level Trace', `ProfileEvent_LogDebug` UInt64 COMMENT 'Number of log messages with level Debug', `ProfileEvent_LogInfo` UInt64 COMMENT 'Number of log messages with level Info', `ProfileEvent_LogWarning` UInt64 COMMENT 'Number of log messages with level Warning', `ProfileEvent_LogError` UInt64 COMMENT 'Number of log messages with level Error', `ProfileEvent_LogFatal` UInt64 COMMENT 'Number of log messages with level Fatal', `ProfileEvent_LoggerElapsedNanoseconds` UInt64 COMMENT 'Cumulative time spend in logging', `ProfileEvent_InterfaceHTTPSendBytes` UInt64 COMMENT 'Number of bytes sent through HTTP interfaces', `ProfileEvent_InterfaceHTTPReceiveBytes` UInt64 COMMENT 'Number of bytes received through HTTP interfaces', `ProfileEvent_InterfaceNativeSendBytes` UInt64 COMMENT 'Number of bytes sent through native interfaces', `ProfileEvent_InterfaceNativeReceiveBytes` UInt64 COMMENT 'Number of bytes received through native interfaces', `ProfileEvent_InterfacePrometheusSendBytes` UInt64 COMMENT 'Number of bytes sent through Prometheus interfaces', `ProfileEvent_InterfacePrometheusReceiveBytes` UInt64 COMMENT 'Number of bytes received through Prometheus interfaces', `ProfileEvent_InterfaceInterserverSendBytes` UInt64 COMMENT 'Number of bytes sent through interserver interfaces', `ProfileEvent_InterfaceInterserverReceiveBytes` UInt64 COMMENT 'Number of bytes received through interserver interfaces', `ProfileEvent_InterfaceMySQLSendBytes` UInt64 COMMENT 'Number of bytes sent through MySQL interfaces', `ProfileEvent_InterfaceMySQLReceiveBytes` UInt64 COMMENT 'Number of bytes received through MySQL interfaces', `ProfileEvent_InterfacePostgreSQLSendBytes` UInt64 COMMENT 'Number of bytes sent through PostgreSQL interfaces', `ProfileEvent_InterfacePostgreSQLReceiveBytes` UInt64 COMMENT 'Number of bytes received through PostgreSQL interfaces', `ProfileEvent_ParallelReplicasUsedCount` UInt64 COMMENT 'Number of replicas used to execute a query with task-based parallel replicas', `ProfileEvent_ParallelReplicasAvailableCount` UInt64 COMMENT 'Number of replicas available to execute a query with task-based parallel replicas', `ProfileEvent_ParallelReplicasUnavailableCount` UInt64 COMMENT 'Number of replicas which was chosen, but found to be unavailable during query execution with task-based parallel replicas', `ProfileEvent_SharedMergeTreeVirtualPartsUpdates` UInt64 COMMENT 'Virtual parts update count', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesByLeader` UInt64 COMMENT 'Virtual parts updates by leader', `ProfileEvent_SharedMergeTreeVirtualPartsUpdateMicroseconds` UInt64 COMMENT 'Virtual parts update microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromZooKeeper` UInt64 COMMENT 'Virtual parts updates count from ZooKeeper', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromZooKeeperMicroseconds` UInt64 COMMENT 'Virtual parts updates from ZooKeeper microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesPeerNotFound` UInt64 COMMENT 'Virtual updates from peer failed because no one found', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromPeer` UInt64 COMMENT 'Virtual parts updates count from peer', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesFromPeerMicroseconds` UInt64 COMMENT 'Virtual parts updates from peer microseconds', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesForMergesOrStatus` UInt64 COMMENT 'Virtual parts updates from non-default background job', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesLeaderFailedElection` UInt64 COMMENT 'Virtual parts updates leader election failed', `ProfileEvent_SharedMergeTreeVirtualPartsUpdatesLeaderSuccessfulElection` UInt64 COMMENT 'Virtual parts updates leader election successful', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentAttempt` UInt64 COMMENT 'How many times we tried to assign merge or mutation', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentFailedWithNothingToDo` UInt64 COMMENT 'How many times we tried to assign merge or mutation and failed because nothing to merge', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentFailedWithConflict` UInt64 COMMENT 'How many times we tried to assign merge or mutation and failed because of conflict in Keeper', `ProfileEvent_SharedMergeTreeMergeMutationAssignmentSuccessful` UInt64 COMMENT 'How many times we tried to assign merge or mutation', `ProfileEvent_SharedMergeTreeMergePartsMovedToOudated` UInt64 COMMENT 'How many parts moved to oudated directory', `ProfileEvent_SharedMergeTreeMergePartsMovedToCondemned` UInt64 COMMENT 'How many parts moved to condemned directory', `ProfileEvent_SharedMergeTreeOutdatedPartsConfirmationRequest` UInt64 COMMENT 'How many ZooKeeper requests were used to config outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsConfirmationInvocations` UInt64 COMMENT 'How many invocations were made to confirm outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsHTTPRequest` UInt64 COMMENT 'How many HTTP requests were send to confirm outdated parts', `ProfileEvent_SharedMergeTreeOutdatedPartsHTTPResponse` UInt64 COMMENT 'How many HTTP responses were send to confirm outdated parts', `ProfileEvent_SharedMergeTreeCondemnedPartsKillRequest` UInt64 COMMENT 'How many ZooKeeper requests were used to remove condemned parts', `ProfileEvent_SharedMergeTreeCondemnedPartsLockConfict` UInt64 COMMENT 'How many times we failed to acquite lock because of conflict', `ProfileEvent_SharedMergeTreeCondemnedPartsRemoved` UInt64 COMMENT 'How many condemned parts were removed', `ProfileEvent_KeeperLogsEntryReadFromLatestCache` UInt64 COMMENT 'Number of log entries in Keeper being read from latest logs cache', `ProfileEvent_KeeperLogsEntryReadFromCommitCache` UInt64 COMMENT 'Number of log entries in Keeper being read from commit logs cache', `ProfileEvent_KeeperLogsEntryReadFromFile` UInt64 COMMENT 'Number of log entries in Keeper being read directly from the changelog file', `ProfileEvent_KeeperLogsPrefetchedEntries` UInt64 COMMENT 'Number of log entries in Keeper being prefetched from the changelog file', `ProfileEvent_StorageConnectionsCreated` UInt64 COMMENT 'Number of created connections for storages', `ProfileEvent_StorageConnectionsReused` UInt64 COMMENT 'Number of reused connections for storages', `ProfileEvent_StorageConnectionsReset` UInt64 COMMENT 'Number of reset connections for storages', `ProfileEvent_StorageConnectionsPreserved` UInt64 COMMENT 'Number of preserved connections for storages', `ProfileEvent_StorageConnectionsExpired` UInt64 COMMENT 'Number of expired connections for storages', `ProfileEvent_StorageConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a connection for storage is failed', `ProfileEvent_StorageConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating connections for storages', `ProfileEvent_DiskConnectionsCreated` UInt64 COMMENT 'Number of created connections for disk', `ProfileEvent_DiskConnectionsReused` UInt64 COMMENT 'Number of reused connections for disk', `ProfileEvent_DiskConnectionsReset` UInt64 COMMENT 'Number of reset connections for disk', `ProfileEvent_DiskConnectionsPreserved` UInt64 COMMENT 'Number of preserved connections for disk', `ProfileEvent_DiskConnectionsExpired` UInt64 COMMENT 'Number of expired connections for disk', `ProfileEvent_DiskConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a connection for disk is failed', `ProfileEvent_DiskConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating connections for disk', `ProfileEvent_HTTPConnectionsCreated` UInt64 COMMENT 'Number of created http connections', `ProfileEvent_HTTPConnectionsReused` UInt64 COMMENT 'Number of reused http connections', `ProfileEvent_HTTPConnectionsReset` UInt64 COMMENT 'Number of reset http connections', `ProfileEvent_HTTPConnectionsPreserved` UInt64 COMMENT 'Number of preserved http connections', `ProfileEvent_HTTPConnectionsExpired` UInt64 COMMENT 'Number of expired http connections', `ProfileEvent_HTTPConnectionsErrors` UInt64 COMMENT 'Number of cases when creation of a http connection failed', `ProfileEvent_HTTPConnectionsElapsedMicroseconds` UInt64 COMMENT 'Total time spend on creating http connections', `ProfileEvent_AddressesDiscovered` UInt64 COMMENT 'Total count of new addresses in dns resolve results for http connections', `ProfileEvent_AddressesExpired` UInt64 COMMENT 'Total count of expired addresses which is no longer presented in dns resolve results for http connections', `ProfileEvent_AddressesMarkedAsFailed` UInt64 COMMENT 'Total count of addresses which has been marked as faulty due to connection errors for http connections', `ProfileEvent_ReadWriteBufferFromHTTPRequestsSent` UInt64 COMMENT 'Number of HTTP requests sent by ReadWriteBufferFromHTTP', `ProfileEvent_ReadWriteBufferFromHTTPBytes` UInt64 COMMENT 'Total size of payload bytes received and sent by ReadWriteBufferFromHTTP. Doesn\'t include HTTP headers.', `ProfileEvent_ConcurrencyControlSlotsGranted` UInt64 COMMENT 'Number of CPU slot granted according to guarantee of 1 thread per query and for queries with setting \'use_concurrency_control\' = 0', `ProfileEvent_ConcurrencyControlSlotsDelayed` UInt64 COMMENT 'Number of CPU slot not granted initially and required to wait for a free CPU slot', `ProfileEvent_ConcurrencyControlSlotsAcquired` UInt64 COMMENT 'Total number of CPU slot acquired', `ProfileEvent_ConcurrencyControlSlotsAcquiredNonCompeting` UInt64 COMMENT 'Total number of noncompeting CPU slot acquired', `ProfileEvent_ConcurrencyControlQueriesDelayed` UInt64 COMMENT 'Total number of CPU slot allocations (queries) that were required to wait for slots to upscale', `ProfileEvent_SharedDatabaseCatalogFailedToApplyState` UInt64 COMMENT 'Number of failures to apply new state in SharedDatabaseCatalog', `ProfileEvent_SharedDatabaseCatalogStateApplicationMicroseconds` UInt64 COMMENT 'Total time spend on application of new state in SharedDatabaseCatalog', `ProfileEvent_GWPAsanAllocateSuccess` UInt64 COMMENT 'Number of successful allocations done by GWPAsan', `ProfileEvent_GWPAsanAllocateFailed` UInt64 COMMENT 'Number of failed allocations done by GWPAsan (i.e. filled pool)', `ProfileEvent_GWPAsanFree` UInt64 COMMENT 'Number of free operations done by GWPAsan', `ProfileEvent_MemoryWorkerRun` UInt64 COMMENT 'Number of runs done by MemoryWorker in background', `ProfileEvent_MemoryWorkerRunElapsedMicroseconds` UInt64 COMMENT 'Total time spent by MemoryWorker for background work', `ProfileEvent_ParquetFetchWaitTimeMicroseconds` UInt64 COMMENT 'Time of waiting fetching parquet data', `CurrentMetric_Query` Int64 COMMENT 'Number of executing queries', `CurrentMetric_Merge` Int64 COMMENT 'Number of executing background merges', `CurrentMetric_MergeParts` Int64 COMMENT 'Number of source parts participating in current background merges', `CurrentMetric_Move` Int64 COMMENT 'Number of currently executing moves', `CurrentMetric_PartMutation` Int64 COMMENT 'Number of mutations (ALTER DELETE/UPDATE)', `CurrentMetric_ReplicatedFetch` Int64 COMMENT 'Number of data parts being fetched from replica', `CurrentMetric_ReplicatedSend` Int64 COMMENT 'Number of data parts being sent to replicas', `CurrentMetric_ReplicatedChecks` Int64 COMMENT 'Number of data parts checking for consistency', `CurrentMetric_BackgroundMergesAndMutationsPoolTask` Int64 COMMENT 'Number of active merges and mutations in an associated background pool', `CurrentMetric_BackgroundMergesAndMutationsPoolSize` Int64 COMMENT 'Limit on number of active merges and mutations in an associated background pool', `CurrentMetric_BackgroundFetchesPoolTask` Int64 COMMENT 'Number of active fetches in an associated background pool', `CurrentMetric_BackgroundFetchesPoolSize` Int64 COMMENT 'Limit on number of simultaneous fetches in an associated background pool', `CurrentMetric_BackgroundCommonPoolTask` Int64 COMMENT 'Number of active tasks in an associated background pool', `CurrentMetric_BackgroundCommonPoolSize` Int64 COMMENT 'Limit on number of tasks in an associated background pool', `CurrentMetric_BackgroundMovePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundProcessingPool for moves', `CurrentMetric_BackgroundMovePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundProcessingPool for moves', `CurrentMetric_BackgroundSchedulePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundSchedulePool. This pool is used for periodic ReplicatedMergeTree tasks, like cleaning old data parts, altering data parts, replica re-initialization, etc.', `CurrentMetric_BackgroundSchedulePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundSchedulePool. This pool is used for periodic ReplicatedMergeTree tasks, like cleaning old data parts, altering data parts, replica re-initialization, etc.', `CurrentMetric_BackgroundBufferFlushSchedulePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundBufferFlushSchedulePool. This pool is used for periodic Buffer flushes', `CurrentMetric_BackgroundBufferFlushSchedulePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundBufferFlushSchedulePool', `CurrentMetric_BackgroundDistributedSchedulePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundDistributedSchedulePool. This pool is used for distributed sends that is done in background.', `CurrentMetric_BackgroundDistributedSchedulePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundDistributedSchedulePool', `CurrentMetric_BackgroundMessageBrokerSchedulePoolTask` Int64 COMMENT 'Number of active tasks in BackgroundMessageBrokerSchedulePool for message streaming', `CurrentMetric_BackgroundMessageBrokerSchedulePoolSize` Int64 COMMENT 'Limit on number of tasks in BackgroundMessageBrokerSchedulePool for message streaming', `CurrentMetric_CacheDictionaryUpdateQueueBatches` Int64 COMMENT 'Number of \'batches\' (a set of keys) in update queue in CacheDictionaries.', `CurrentMetric_CacheDictionaryUpdateQueueKeys` Int64 COMMENT 'Exact number of keys in update queue in CacheDictionaries.', `CurrentMetric_DiskSpaceReservedForMerge` Int64 COMMENT 'Disk space reserved for currently running background merges. It is slightly more than the total size of currently merging parts.', `CurrentMetric_DistributedSend` Int64 COMMENT 'Number of connections to remote servers sending data that was INSERTed into Distributed tables. Both synchronous and asynchronous mode.', `CurrentMetric_QueryPreempted` Int64 COMMENT 'Number of queries that are stopped and waiting due to \'priority\' setting.', `CurrentMetric_TCPConnection` Int64 COMMENT 'Number of connections to TCP server (clients with native interface), also included server-server distributed query connections', `CurrentMetric_MySQLConnection` Int64 COMMENT 'Number of client connections using MySQL protocol', `CurrentMetric_HTTPConnection` Int64 COMMENT 'Number of connections to HTTP server', `CurrentMetric_InterserverConnection` Int64 COMMENT 'Number of connections from other replicas to fetch parts', `CurrentMetric_PostgreSQLConnection` Int64 COMMENT 'Number of client connections using PostgreSQL protocol', `CurrentMetric_OpenFileForRead` Int64 COMMENT 'Number of files open for reading', `CurrentMetric_OpenFileForWrite` Int64 COMMENT 'Number of files open for writing', `CurrentMetric_Compressing` Int64 COMMENT 'Number of compress operations using internal compression codecs', `CurrentMetric_Decompressing` Int64 COMMENT 'Number of decompress operations using internal compression codecs', `CurrentMetric_ParallelCompressedWriteBufferThreads` Int64 COMMENT 'Number of threads in all instances of ParallelCompressedWriteBuffer - these threads are doing parallel compression and writing', `CurrentMetric_ParallelCompressedWriteBufferWait` Int64 COMMENT 'Number of threads in all instances of ParallelCompressedWriteBuffer that are currently waiting for buffer to become available for writing', `CurrentMetric_TotalTemporaryFiles` Int64 COMMENT 'Number of temporary files created', `CurrentMetric_TemporaryFilesForSort` Int64 COMMENT 'Number of temporary files created for external sorting', `CurrentMetric_TemporaryFilesForAggregation` Int64 COMMENT 'Number of temporary files created for external aggregation', `CurrentMetric_TemporaryFilesForJoin` Int64 COMMENT 'Number of temporary files created for JOIN', `CurrentMetric_TemporaryFilesForMerge` Int64 COMMENT 'Number of temporary files for vertical merge', `CurrentMetric_TemporaryFilesUnknown` Int64 COMMENT 'Number of temporary files created without known purpose', `CurrentMetric_Read` Int64 COMMENT 'Number of read (read, pread, io_getevents, etc.) syscalls in fly', `CurrentMetric_RemoteRead` Int64 COMMENT 'Number of read with remote reader in fly', `CurrentMetric_Write` Int64 COMMENT 'Number of write (write, pwrite, io_getevents, etc.) syscalls in fly', `CurrentMetric_NetworkReceive` Int64 COMMENT 'Number of threads receiving data from network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `CurrentMetric_NetworkSend` Int64 COMMENT 'Number of threads sending data to network. Only ClickHouse-related network interaction is included, not by 3rd party libraries.', `CurrentMetric_SendScalars` Int64 COMMENT 'Number of connections that are sending data for scalars to remote servers.', `CurrentMetric_SendExternalTables` Int64 COMMENT 'Number of connections that are sending data for external tables to remote servers. External tables are used to implement GLOBAL IN and GLOBAL JOIN operators with distributed subqueries.', `CurrentMetric_QueryThread` Int64 COMMENT 'Number of query processing threads', `CurrentMetric_ReadonlyReplica` Int64 COMMENT 'Number of Replicated tables that are currently in readonly state due to re-initialization after ZooKeeper session loss or due to startup without ZooKeeper configured.', `CurrentMetric_MemoryTracking` Int64 COMMENT 'Total amount of memory (bytes) allocated by the server.', `CurrentMetric_MemoryTrackingUncorrected` Int64 COMMENT 'Total amount of memory (bytes) allocated by the server not corrected by RSS.', `CurrentMetric_MergesMutationsMemoryTracking` Int64 COMMENT 'Total amount of memory (bytes) allocated by background tasks (merges and mutations).', `CurrentMetric_EphemeralNode` Int64 COMMENT 'Number of ephemeral nodes hold in ZooKeeper.', `CurrentMetric_ZooKeeperSession` Int64 COMMENT 'Number of sessions (connections) to ZooKeeper. Should be no more than one, because using more than one connection to ZooKeeper may lead to bugs due to lack of linearizability (stale reads) that ZooKeeper consistency model allows.', `CurrentMetric_ZooKeeperWatch` Int64 COMMENT 'Number of watches (event subscriptions) in ZooKeeper.', `CurrentMetric_ZooKeeperRequest` Int64 COMMENT 'Number of requests to ZooKeeper in fly.', `CurrentMetric_DelayedInserts` Int64 COMMENT 'Number of INSERT queries that are throttled due to high number of active data parts for partition in a MergeTree table.', `CurrentMetric_ContextLockWait` Int64 COMMENT 'Number of threads waiting for lock in Context. This is global lock.', `CurrentMetric_StorageBufferRows` Int64 COMMENT 'Number of rows in buffers of Buffer tables', `CurrentMetric_StorageBufferBytes` Int64 COMMENT 'Number of bytes in buffers of Buffer tables', `CurrentMetric_DictCacheRequests` Int64 COMMENT 'Number of requests in fly to data sources of dictionaries of cache type.', `CurrentMetric_Revision` Int64 COMMENT 'Revision of the server. It is a number incremented for every release or release candidate except patch releases.', `CurrentMetric_VersionInteger` Int64 COMMENT 'Version of the server in a single integer number in base-1000. For example, version 11.22.33 is translated to 11022033.', `CurrentMetric_RWLockWaitingReaders` Int64 COMMENT 'Number of threads waiting for read on a table RWLock.', `CurrentMetric_RWLockWaitingWriters` Int64 COMMENT 'Number of threads waiting for write on a table RWLock.', `CurrentMetric_RWLockActiveReaders` Int64 COMMENT 'Number of threads holding read lock in a table RWLock.', `CurrentMetric_RWLockActiveWriters` Int64 COMMENT 'Number of threads holding write lock in a table RWLock.', `CurrentMetric_GlobalThread` Int64 COMMENT 'Number of threads in global thread pool.', `CurrentMetric_GlobalThreadActive` Int64 COMMENT 'Number of threads in global thread pool running a task.', `CurrentMetric_GlobalThreadScheduled` Int64 COMMENT 'Number of queued or active jobs in global thread pool.', `CurrentMetric_LocalThread` Int64 COMMENT 'Obsolete. Number of threads in local thread pools. The threads in local thread pools are taken from the global thread pool.', `CurrentMetric_LocalThreadActive` Int64 COMMENT 'Obsolete. Number of threads in local thread pools running a task.', `CurrentMetric_LocalThreadScheduled` Int64 COMMENT 'Obsolete. Number of queued or active jobs in local thread pools.', `CurrentMetric_MergeTreeDataSelectExecutorThreads` Int64 COMMENT 'Number of threads in the MergeTreeDataSelectExecutor thread pool.', `CurrentMetric_MergeTreeDataSelectExecutorThreadsActive` Int64 COMMENT 'Number of threads in the MergeTreeDataSelectExecutor thread pool running a task.', `CurrentMetric_MergeTreeDataSelectExecutorThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the MergeTreeDataSelectExecutor thread pool.', `CurrentMetric_BackupsThreads` Int64 COMMENT 'Number of threads in the thread pool for BACKUP.', `CurrentMetric_BackupsThreadsActive` Int64 COMMENT 'Number of threads in thread pool for BACKUP running a task.', `CurrentMetric_BackupsThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs for BACKUP.', `CurrentMetric_RestoreThreads` Int64 COMMENT 'Number of threads in the thread pool for RESTORE.', `CurrentMetric_RestoreThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for RESTORE running a task.', `CurrentMetric_RestoreThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs for RESTORE.', `CurrentMetric_MarksLoaderThreads` Int64 COMMENT 'Number of threads in thread pool for loading marks.', `CurrentMetric_MarksLoaderThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for loading marks running a task.', `CurrentMetric_MarksLoaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for loading marks.', `CurrentMetric_IOPrefetchThreads` Int64 COMMENT 'Number of threads in the IO prefetch thread pool.', `CurrentMetric_IOPrefetchThreadsActive` Int64 COMMENT 'Number of threads in the IO prefetch thread pool running a task.', `CurrentMetric_IOPrefetchThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the IO prefetch thread pool.', `CurrentMetric_IOWriterThreads` Int64 COMMENT 'Number of threads in the IO writer thread pool.', `CurrentMetric_IOWriterThreadsActive` Int64 COMMENT 'Number of threads in the IO writer thread pool running a task.', `CurrentMetric_IOWriterThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the IO writer thread pool.', `CurrentMetric_IOThreads` Int64 COMMENT 'Number of threads in the IO thread pool.', `CurrentMetric_IOThreadsActive` Int64 COMMENT 'Number of threads in the IO thread pool running a task.', `CurrentMetric_IOThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the IO thread pool.', `CurrentMetric_CompressionThread` Int64 COMMENT 'Number of threads in compression thread pools.', `CurrentMetric_CompressionThreadActive` Int64 COMMENT 'Number of threads in compression thread pools running a task.', `CurrentMetric_CompressionThreadScheduled` Int64 COMMENT 'Number of queued or active jobs in compression thread pools.', `CurrentMetric_ThreadPoolRemoteFSReaderThreads` Int64 COMMENT 'Number of threads in the thread pool for remote_filesystem_read_method=threadpool.', `CurrentMetric_ThreadPoolRemoteFSReaderThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for remote_filesystem_read_method=threadpool running a task.', `CurrentMetric_ThreadPoolRemoteFSReaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for remote_filesystem_read_method=threadpool.', `CurrentMetric_ThreadPoolFSReaderThreads` Int64 COMMENT 'Number of threads in the thread pool for local_filesystem_read_method=threadpool.', `CurrentMetric_ThreadPoolFSReaderThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for local_filesystem_read_method=threadpool running a task.', `CurrentMetric_ThreadPoolFSReaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for local_filesystem_read_method=threadpool.', `CurrentMetric_BackupsIOThreads` Int64 COMMENT 'Number of threads in the BackupsIO thread pool.', `CurrentMetric_BackupsIOThreadsActive` Int64 COMMENT 'Number of threads in the BackupsIO thread pool running a task.', `CurrentMetric_BackupsIOThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the BackupsIO thread pool.', `CurrentMetric_DiskObjectStorageAsyncThreads` Int64 COMMENT 'Obsolete metric, shows nothing.', `CurrentMetric_DiskObjectStorageAsyncThreadsActive` Int64 COMMENT 'Obsolete metric, shows nothing.', `CurrentMetric_StorageHiveThreads` Int64 COMMENT 'Number of threads in the StorageHive thread pool.', `CurrentMetric_StorageHiveThreadsActive` Int64 COMMENT 'Number of threads in the StorageHive thread pool running a task.', `CurrentMetric_StorageHiveThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the StorageHive thread pool.', `CurrentMetric_TablesLoaderBackgroundThreads` Int64 COMMENT 'Number of threads in the tables loader background thread pool.', `CurrentMetric_TablesLoaderBackgroundThreadsActive` Int64 COMMENT 'Number of threads in the tables loader background thread pool running a task.', `CurrentMetric_TablesLoaderBackgroundThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the tables loader background thread pool.', `CurrentMetric_TablesLoaderForegroundThreads` Int64 COMMENT 'Number of threads in the tables loader foreground thread pool.', `CurrentMetric_TablesLoaderForegroundThreadsActive` Int64 COMMENT 'Number of threads in the tables loader foreground thread pool running a task.', `CurrentMetric_TablesLoaderForegroundThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the tables loader foreground thread pool.', `CurrentMetric_DatabaseOnDiskThreads` Int64 COMMENT 'Number of threads in the DatabaseOnDisk thread pool.', `CurrentMetric_DatabaseOnDiskThreadsActive` Int64 COMMENT 'Number of threads in the DatabaseOnDisk thread pool running a task.', `CurrentMetric_DatabaseOnDiskThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DatabaseOnDisk thread pool.', `CurrentMetric_DatabaseBackupThreads` Int64 COMMENT 'Number of threads in the DatabaseBackup thread pool.', `CurrentMetric_DatabaseBackupThreadsActive` Int64 COMMENT 'Number of threads in the DatabaseBackup thread pool running a task.', `CurrentMetric_DatabaseBackupThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DatabaseBackup thread pool.', `CurrentMetric_DatabaseCatalogThreads` Int64 COMMENT 'Number of threads in the DatabaseCatalog thread pool.', `CurrentMetric_DatabaseCatalogThreadsActive` Int64 COMMENT 'Number of threads in the DatabaseCatalog thread pool running a task.', `CurrentMetric_DatabaseCatalogThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DatabaseCatalog thread pool.', `CurrentMetric_DestroyAggregatesThreads` Int64 COMMENT 'Number of threads in the thread pool for destroy aggregate states.', `CurrentMetric_DestroyAggregatesThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for destroy aggregate states running a task.', `CurrentMetric_DestroyAggregatesThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for destroy aggregate states.', `CurrentMetric_ConcurrentHashJoinPoolThreads` Int64 COMMENT 'Number of threads in the thread pool for concurrent hash join.', `CurrentMetric_ConcurrentHashJoinPoolThreadsActive` Int64 COMMENT 'Number of threads in the thread pool for concurrent hash join running a task.', `CurrentMetric_ConcurrentHashJoinPoolThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool for concurrent hash join.', `CurrentMetric_HashedDictionaryThreads` Int64 COMMENT 'Number of threads in the HashedDictionary thread pool.', `CurrentMetric_HashedDictionaryThreadsActive` Int64 COMMENT 'Number of threads in the HashedDictionary thread pool running a task.', `CurrentMetric_HashedDictionaryThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the HashedDictionary thread pool.', `CurrentMetric_CacheDictionaryThreads` Int64 COMMENT 'Number of threads in the CacheDictionary thread pool.', `CurrentMetric_CacheDictionaryThreadsActive` Int64 COMMENT 'Number of threads in the CacheDictionary thread pool running a task.', `CurrentMetric_CacheDictionaryThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the CacheDictionary thread pool.', `CurrentMetric_ParallelFormattingOutputFormatThreads` Int64 COMMENT 'Number of threads in the ParallelFormattingOutputFormatThreads thread pool.', `CurrentMetric_ParallelFormattingOutputFormatThreadsActive` Int64 COMMENT 'Number of threads in the ParallelFormattingOutputFormatThreads thread pool running a task.', `CurrentMetric_ParallelFormattingOutputFormatThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the ParallelFormattingOutputFormatThreads thread pool.', `CurrentMetric_ParallelParsingInputFormatThreads` Int64 COMMENT 'Number of threads in the ParallelParsingInputFormat thread pool.', `CurrentMetric_ParallelParsingInputFormatThreadsActive` Int64 COMMENT 'Number of threads in the ParallelParsingInputFormat thread pool running a task.', `CurrentMetric_ParallelParsingInputFormatThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the ParallelParsingInputFormat thread pool.', `CurrentMetric_MergeTreeBackgroundExecutorThreads` Int64 COMMENT 'Number of threads in the MergeTreeBackgroundExecutor thread pool.', `CurrentMetric_MergeTreeBackgroundExecutorThreadsActive` Int64 COMMENT 'Number of threads in the MergeTreeBackgroundExecutor thread pool running a task.', `CurrentMetric_MergeTreeBackgroundExecutorThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the MergeTreeBackgroundExecutor thread pool.', `CurrentMetric_AsynchronousInsertThreads` Int64 COMMENT 'Number of threads in the AsynchronousInsert thread pool.', `CurrentMetric_AsynchronousInsertThreadsActive` Int64 COMMENT 'Number of threads in the AsynchronousInsert thread pool running a task.', `CurrentMetric_AsynchronousInsertThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the AsynchronousInsert thread pool.', `CurrentMetric_AsynchronousInsertQueueSize` Int64 COMMENT 'Number of pending tasks in the AsynchronousInsert queue.', `CurrentMetric_AsynchronousInsertQueueBytes` Int64 COMMENT 'Number of pending bytes in the AsynchronousInsert queue.', `CurrentMetric_StartupSystemTablesThreads` Int64 COMMENT 'Number of threads in the StartupSystemTables thread pool.', `CurrentMetric_StartupSystemTablesThreadsActive` Int64 COMMENT 'Number of threads in the StartupSystemTables thread pool running a task.', `CurrentMetric_StartupSystemTablesThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the StartupSystemTables thread pool.', `CurrentMetric_AggregatorThreads` Int64 COMMENT 'Number of threads in the Aggregator thread pool.', `CurrentMetric_AggregatorThreadsActive` Int64 COMMENT 'Number of threads in the Aggregator thread pool running a task.', `CurrentMetric_AggregatorThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the Aggregator thread pool.', `CurrentMetric_DDLWorkerThreads` Int64 COMMENT 'Number of threads in the DDLWorker thread pool for ON CLUSTER queries.', `CurrentMetric_DDLWorkerThreadsActive` Int64 COMMENT 'Number of threads in the DDLWORKER thread pool for ON CLUSTER queries running a task.', `CurrentMetric_DDLWorkerThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DDLWORKER thread pool for ON CLUSTER queries.', `CurrentMetric_StorageDistributedThreads` Int64 COMMENT 'Number of threads in the StorageDistributed thread pool.', `CurrentMetric_StorageDistributedThreadsActive` Int64 COMMENT 'Number of threads in the StorageDistributed thread pool running a task.', `CurrentMetric_StorageDistributedThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the StorageDistributed thread pool.', `CurrentMetric_DistributedInsertThreads` Int64 COMMENT 'Number of threads used for INSERT into Distributed.', `CurrentMetric_DistributedInsertThreadsActive` Int64 COMMENT 'Number of threads used for INSERT into Distributed running a task.', `CurrentMetric_DistributedInsertThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs used for INSERT into Distributed.', `CurrentMetric_StorageS3Threads` Int64 COMMENT 'Number of threads in the StorageS3 thread pool.', `CurrentMetric_StorageS3ThreadsActive` Int64 COMMENT 'Number of threads in the StorageS3 thread pool running a task.', `CurrentMetric_StorageS3ThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the StorageS3 thread pool.', `CurrentMetric_ObjectStorageS3Threads` Int64 COMMENT 'Number of threads in the S3ObjectStorage thread pool.', `CurrentMetric_ObjectStorageS3ThreadsActive` Int64 COMMENT 'Number of threads in the S3ObjectStorage thread pool running a task.', `CurrentMetric_ObjectStorageS3ThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the S3ObjectStorage thread pool.', `CurrentMetric_StorageObjectStorageThreads` Int64 COMMENT 'Number of threads in the remote table engines thread pools.', `CurrentMetric_StorageObjectStorageThreadsActive` Int64 COMMENT 'Number of threads in the remote table engines thread pool running a task.', `CurrentMetric_StorageObjectStorageThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in remote table engines thread pool.', `CurrentMetric_ObjectStorageAzureThreads` Int64 COMMENT 'Number of threads in the AzureObjectStorage thread pool.', `CurrentMetric_ObjectStorageAzureThreadsActive` Int64 COMMENT 'Number of threads in the AzureObjectStorage thread pool running a task.', `CurrentMetric_ObjectStorageAzureThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the AzureObjectStorage thread pool.', `CurrentMetric_BuildVectorSimilarityIndexThreads` Int64 COMMENT 'Number of threads in the build vector similarity index thread pool.', `CurrentMetric_BuildVectorSimilarityIndexThreadsActive` Int64 COMMENT 'Number of threads in the build vector similarity index thread pool running a task.', `CurrentMetric_BuildVectorSimilarityIndexThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the build vector similarity index thread pool.', `CurrentMetric_ObjectStorageQueueRegisteredServers` Int64 COMMENT 'Number of registered servers in StorageS3(Azure)Queue', `CurrentMetric_IcebergCatalogThreads` Int64 COMMENT 'Number of threads in the IcebergCatalog thread pool.', `CurrentMetric_IcebergCatalogThreadsActive` Int64 COMMENT 'Number of threads in the IcebergCatalog thread pool running a task.', `CurrentMetric_IcebergCatalogThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the IcebergCatalog thread pool.', `CurrentMetric_ParallelWithQueryThreads` Int64 COMMENT 'Number of threads in the threadpool for processing PARALLEL WITH queries.', `CurrentMetric_ParallelWithQueryActiveThreads` Int64 COMMENT 'Number of active threads in the threadpool for processing PARALLEL WITH queries.', `CurrentMetric_ParallelWithQueryScheduledThreads` Int64 COMMENT 'Number of queued or active jobs in the threadpool for processing PARALLEL WITH queries.', `CurrentMetric_DiskPlainRewritableAzureDirectoryMapSize` Int64 COMMENT 'Number of local-to-remote path entries in the \'plain_rewritable\' in-memory map for AzureObjectStorage.', `CurrentMetric_DiskPlainRewritableAzureFileCount` Int64 COMMENT 'Number of file entries in the \'plain_rewritable\' in-memory map for AzureObjectStorage.', `CurrentMetric_DiskPlainRewritableAzureUniqueFileNamesCount` Int64 COMMENT 'Number of unique file name entries in the \'plain_rewritable\' in-memory map for AzureObjectStorage.', `CurrentMetric_DiskPlainRewritableLocalDirectoryMapSize` Int64 COMMENT 'Number of local-to-remote path entries in the \'plain_rewritable\' in-memory map for LocalObjectStorage.', `CurrentMetric_DiskPlainRewritableLocalFileCount` Int64 COMMENT 'Number of file entries in the \'plain_rewritable\' in-memory map for LocalObjectStorage.', `CurrentMetric_DiskPlainRewritableLocalUniqueFileNamesCount` Int64 COMMENT 'Number of unique file name entries in the \'plain_rewritable\' in-memory map for LocalObjectStorage.', `CurrentMetric_DiskPlainRewritableS3DirectoryMapSize` Int64 COMMENT 'Number of local-to-remote path entries in the \'plain_rewritable\' in-memory map for S3ObjectStorage.', `CurrentMetric_DiskPlainRewritableS3FileCount` Int64 COMMENT 'Number of file entries in the \'plain_rewritable\' in-memory map for S3ObjectStorage.', `CurrentMetric_DiskPlainRewritableS3UniqueFileNamesCount` Int64 COMMENT 'Number of unique file name entries in the \'plain_rewritable\' in-memory map for S3ObjectStorage.', `CurrentMetric_MergeTreeFetchPartitionThreads` Int64 COMMENT 'Number of threads for ALTER TABLE FETCH PARTITION', `CurrentMetric_MergeTreeFetchPartitionThreadsActive` Int64 COMMENT 'Number of threads for ALTER TABLE FETCH PARTITION fetching part', `CurrentMetric_MergeTreeFetchPartitionThreadsScheduled` Int64 COMMENT 'Number of queued or active part fetches in ALTER TABLE FETCH PARTITION', `CurrentMetric_MergeTreePartsLoaderThreads` Int64 COMMENT 'Number of threads in the MergeTree parts loader thread pool.', `CurrentMetric_MergeTreePartsLoaderThreadsActive` Int64 COMMENT 'Number of threads in the MergeTree parts loader thread pool running a task.', `CurrentMetric_MergeTreePartsLoaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the MergeTree parts loader thread pool.', `CurrentMetric_MergeTreeOutdatedPartsLoaderThreads` Int64 COMMENT 'Number of threads in the threadpool for loading Outdated data parts.', `CurrentMetric_MergeTreeOutdatedPartsLoaderThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for loading Outdated data parts.', `CurrentMetric_MergeTreeOutdatedPartsLoaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for loading Outdated data parts.', `CurrentMetric_MergeTreeUnexpectedPartsLoaderThreads` Int64 COMMENT 'Number of threads in the threadpool for loading Unexpected data parts.', `CurrentMetric_MergeTreeUnexpectedPartsLoaderThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for loading Unexpected data parts.', `CurrentMetric_MergeTreeUnexpectedPartsLoaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for loading Unexpected data parts.', `CurrentMetric_MergeTreePartsCleanerThreads` Int64 COMMENT 'Number of threads in the MergeTree parts cleaner thread pool.', `CurrentMetric_MergeTreePartsCleanerThreadsActive` Int64 COMMENT 'Number of threads in the MergeTree parts cleaner thread pool running a task.', `CurrentMetric_MergeTreePartsCleanerThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the MergeTree parts cleaner thread pool.', `CurrentMetric_DatabaseReplicatedCreateTablesThreads` Int64 COMMENT 'Number of threads in the threadpool for table creation in DatabaseReplicated.', `CurrentMetric_DatabaseReplicatedCreateTablesThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for table creation in DatabaseReplicated.', `CurrentMetric_DatabaseReplicatedCreateTablesThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for table creation in DatabaseReplicated.', `CurrentMetric_IDiskCopierThreads` Int64 COMMENT 'Number of threads for copying data between disks of different types.', `CurrentMetric_IDiskCopierThreadsActive` Int64 COMMENT 'Number of threads for copying data between disks of different types running a task.', `CurrentMetric_IDiskCopierThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs for copying data between disks of different types.', `CurrentMetric_SystemReplicasThreads` Int64 COMMENT 'Number of threads in the system.replicas thread pool.', `CurrentMetric_SystemReplicasThreadsActive` Int64 COMMENT 'Number of threads in the system.replicas thread pool running a task.', `CurrentMetric_SystemReplicasThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the system.replicas thread pool.', `CurrentMetric_RestartReplicaThreads` Int64 COMMENT 'Number of threads in the RESTART REPLICA thread pool.', `CurrentMetric_RestartReplicaThreadsActive` Int64 COMMENT 'Number of threads in the RESTART REPLICA thread pool running a task.', `CurrentMetric_RestartReplicaThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the RESTART REPLICA thread pool.', `CurrentMetric_QueryPipelineExecutorThreads` Int64 COMMENT 'Number of threads in the PipelineExecutor thread pool.', `CurrentMetric_QueryPipelineExecutorThreadsActive` Int64 COMMENT 'Number of threads in the PipelineExecutor thread pool running a task.', `CurrentMetric_QueryPipelineExecutorThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the PipelineExecutor thread pool.', `CurrentMetric_ParquetDecoderThreads` Int64 COMMENT 'Number of threads in the ParquetBlockInputFormat thread pool.', `CurrentMetric_ParquetDecoderThreadsActive` Int64 COMMENT 'Number of threads in the ParquetBlockInputFormat thread pool running a task.', `CurrentMetric_ParquetDecoderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the ParquetBlockInputFormat thread pool.', `CurrentMetric_ParquetDecoderIOThreads` Int64 COMMENT 'Number of threads in the ParquetBlockInputFormat io thread pool.', `CurrentMetric_ParquetDecoderIOThreadsActive` Int64 COMMENT 'Number of threads in the ParquetBlockInputFormat io thread pool running a task.', `CurrentMetric_ParquetDecoderIOThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the ParquetBlockInputFormat io thread pool.', `CurrentMetric_ParquetEncoderThreads` Int64 COMMENT 'Number of threads in ParquetBlockOutputFormat thread pool.', `CurrentMetric_ParquetEncoderThreadsActive` Int64 COMMENT 'Number of threads in ParquetBlockOutputFormat thread pool running a task.', `CurrentMetric_ParquetEncoderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in ParquetBlockOutputFormat thread pool.', `CurrentMetric_MergeTreeSubcolumnsReaderThreads` Int64 COMMENT 'Number of threads in the thread pool used for subcolumns reading in MergeTree.', `CurrentMetric_MergeTreeSubcolumnsReaderThreadsActive` Int64 COMMENT 'Number of threads in the thread pool used for subcolumns reading in MergeTree running a task.', `CurrentMetric_MergeTreeSubcolumnsReaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the thread pool used for subcolumns reading in MergeTree.', `CurrentMetric_DWARFReaderThreads` Int64 COMMENT 'Number of threads in the DWARFBlockInputFormat thread pool.', `CurrentMetric_DWARFReaderThreadsActive` Int64 COMMENT 'Number of threads in the DWARFBlockInputFormat thread pool running a task.', `CurrentMetric_DWARFReaderThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the DWARFBlockInputFormat thread pool.', `CurrentMetric_OutdatedPartsLoadingThreads` Int64 COMMENT 'Number of threads in the threadpool for loading Outdated data parts.', `CurrentMetric_OutdatedPartsLoadingThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for loading Outdated data parts.', `CurrentMetric_OutdatedPartsLoadingThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for loading Outdated data parts.', `CurrentMetric_PolygonDictionaryThreads` Int64 COMMENT 'Number of threads in the threadpool for polygon dictionaries.', `CurrentMetric_PolygonDictionaryThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for polygon dictionaries.', `CurrentMetric_PolygonDictionaryThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for polygon dictionaries.', `CurrentMetric_DistributedBytesToInsert` Int64 COMMENT 'Number of pending bytes to process for asynchronous insertion into Distributed tables. Number of bytes for every shard is summed.', `CurrentMetric_BrokenDistributedBytesToInsert` Int64 COMMENT 'Number of bytes for asynchronous insertion into Distributed tables that has been marked as broken. Number of bytes for every shard is summed.', `CurrentMetric_DistributedFilesToInsert` Int64 COMMENT 'Number of pending files to process for asynchronous insertion into Distributed tables. Number of files for every shard is summed.', `CurrentMetric_BrokenDistributedFilesToInsert` Int64 COMMENT 'Number of files for asynchronous insertion into Distributed tables that has been marked as broken. Number of files for every shard is summed.', `CurrentMetric_TablesToDropQueueSize` Int64 COMMENT 'Number of dropped tables, that are waiting for background data removal.', `CurrentMetric_MaxDDLEntryID` Int64 COMMENT 'Max processed DDL entry of DDLWorker.', `CurrentMetric_MaxPushedDDLEntryID` Int64 COMMENT 'Max DDL entry of DDLWorker that pushed to zookeeper.', `CurrentMetric_PartsTemporary` Int64 COMMENT 'The part is generating now, it is not in data_parts list.', `CurrentMetric_PartsPreCommitted` Int64 COMMENT 'Deprecated. See PartsPreActive.', `CurrentMetric_PartsCommitted` Int64 COMMENT 'Deprecated. See PartsActive.', `CurrentMetric_PartsPreActive` Int64 COMMENT 'The part is in data_parts, but not used for SELECTs.', `CurrentMetric_PartsActive` Int64 COMMENT 'Active data part, used by current and upcoming SELECTs.', `CurrentMetric_AttachedDatabase` Int64 COMMENT 'Active databases.', `CurrentMetric_AttachedTable` Int64 COMMENT 'Active tables.', `CurrentMetric_AttachedReplicatedTable` Int64 COMMENT 'Active replicated tables.', `CurrentMetric_AttachedView` Int64 COMMENT 'Active views.', `CurrentMetric_AttachedDictionary` Int64 COMMENT 'Active dictionaries.', `CurrentMetric_PartsOutdated` Int64 COMMENT 'Not active data part, but could be used by only current SELECTs, could be deleted after SELECTs finishes.', `CurrentMetric_PartsDeleting` Int64 COMMENT 'Not active data part with identity refcounter, it is deleting right now by a cleaner.', `CurrentMetric_PartsDeleteOnDestroy` Int64 COMMENT 'Part was moved to another disk and should be deleted in own destructor.', `CurrentMetric_PartsWide` Int64 COMMENT 'Wide parts.', `CurrentMetric_PartsCompact` Int64 COMMENT 'Compact parts.', `CurrentMetric_MMappedFiles` Int64 COMMENT 'Total number of mmapped files.', `CurrentMetric_MMappedFileBytes` Int64 COMMENT 'Sum size of mmapped file regions.', `CurrentMetric_AsynchronousReadWait` Int64 COMMENT 'Number of threads waiting for asynchronous read.', `CurrentMetric_PendingAsyncInsert` Int64 COMMENT 'Number of asynchronous inserts that are waiting for flush.', `CurrentMetric_KafkaConsumers` Int64 COMMENT 'Number of active Kafka consumers', `CurrentMetric_KafkaConsumersWithAssignment` Int64 COMMENT 'Number of active Kafka consumers which have some partitions assigned.', `CurrentMetric_KafkaProducers` Int64 COMMENT 'Number of active Kafka producer created', `CurrentMetric_KafkaLibrdkafkaThreads` Int64 COMMENT 'Number of active librdkafka threads', `CurrentMetric_KafkaBackgroundReads` Int64 COMMENT 'Number of background reads currently working (populating materialized views from Kafka)', `CurrentMetric_KafkaConsumersInUse` Int64 COMMENT 'Number of consumers which are currently used by direct or background reads', `CurrentMetric_KafkaWrites` Int64 COMMENT 'Number of currently running inserts to Kafka', `CurrentMetric_KafkaAssignedPartitions` Int64 COMMENT 'Number of partitions Kafka tables currently assigned to', `CurrentMetric_FilesystemCacheReadBuffers` Int64 COMMENT 'Number of active cache buffers', `CurrentMetric_CacheFileSegments` Int64 COMMENT 'Number of existing cache file segments', `CurrentMetric_CacheDetachedFileSegments` Int64 COMMENT 'Number of existing detached cache file segments', `CurrentMetric_FilesystemCacheSize` Int64 COMMENT 'Filesystem cache size in bytes', `CurrentMetric_FilesystemCacheSizeLimit` Int64 COMMENT 'Filesystem cache size limit in bytes', `CurrentMetric_FilesystemCacheElements` Int64 COMMENT 'Filesystem cache elements (file segments)', `CurrentMetric_FilesystemCacheDownloadQueueElements` Int64 COMMENT 'Filesystem cache elements in download queue', `CurrentMetric_FilesystemCacheDelayedCleanupElements` Int64 COMMENT 'Filesystem cache elements in background cleanup queue', `CurrentMetric_FilesystemCacheHoldFileSegments` Int64 COMMENT 'Filesystem cache file segment which are currently hold as unreleasable', `CurrentMetric_AsyncInsertCacheSize` Int64 COMMENT 'Number of async insert hash id in cache', `CurrentMetric_SkippingIndexCacheSize` Int64 COMMENT 'Size of the skipping index cache in bytes', `CurrentMetric_S3Requests` Int64 COMMENT 'S3 requests count', `CurrentMetric_KeeperAliveConnections` Int64 COMMENT 'Number of alive connections', `CurrentMetric_KeeperOutstandingRequests` Int64 COMMENT 'Number of outstanding requests', `CurrentMetric_ThreadsInOvercommitTracker` Int64 COMMENT 'Number of waiting threads inside of OvercommitTracker', `CurrentMetric_IOUringPendingEvents` Int64 COMMENT 'Number of io_uring SQEs waiting to be submitted', `CurrentMetric_IOUringInFlightEvents` Int64 COMMENT 'Number of io_uring SQEs in flight', `CurrentMetric_ReadTaskRequestsSent` Int64 COMMENT 'The current number of callback requests in flight from the remote server back to the initiator server to choose the read task (for s3Cluster table function and similar). Measured on the remote server side.', `CurrentMetric_MergeTreeReadTaskRequestsSent` Int64 COMMENT 'The current number of callback requests in flight from the remote server back to the initiator server to choose the read task (for MergeTree tables). Measured on the remote server side.', `CurrentMetric_MergeTreeAllRangesAnnouncementsSent` Int64 COMMENT 'The current number of announcement being sent in flight from the remote server to the initiator server about the set of data parts (for MergeTree tables). Measured on the remote server side.', `CurrentMetric_CreatedTimersInQueryProfiler` Int64 COMMENT 'Number of Created thread local timers in QueryProfiler', `CurrentMetric_ActiveTimersInQueryProfiler` Int64 COMMENT 'Number of Active thread local timers in QueryProfiler', `CurrentMetric_RefreshableViews` Int64 COMMENT 'Number materialized views with periodic refreshing (REFRESH)', `CurrentMetric_RefreshingViews` Int64 COMMENT 'Number of materialized views currently executing a refresh', `CurrentMetric_StorageBufferFlushThreads` Int64 COMMENT 'Number of threads for background flushes in StorageBuffer', `CurrentMetric_StorageBufferFlushThreadsActive` Int64 COMMENT 'Number of threads for background flushes in StorageBuffer running a task', `CurrentMetric_StorageBufferFlushThreadsScheduled` Int64 COMMENT 'Number of queued or active threads for background flushes in StorageBuffer', `CurrentMetric_SharedMergeTreeThreads` Int64 COMMENT 'Number of threads in the thread pools in internals of SharedMergeTree', `CurrentMetric_SharedMergeTreeThreadsActive` Int64 COMMENT 'Number of threads in the thread pools in internals of SharedMergeTree running a task', `CurrentMetric_SharedMergeTreeThreadsScheduled` Int64 COMMENT 'Number of queued or active threads in the thread pools in internals of SharedMergeTree', `CurrentMetric_SharedMergeTreeFetch` Int64 COMMENT 'Number of fetches in progress', `CurrentMetric_CacheWarmerBytesInProgress` Int64 COMMENT 'Total size of remote file segments waiting to be asynchronously loaded into filesystem cache.', `CurrentMetric_DistrCacheOpenedConnections` Int64 COMMENT 'Number of open connections to Distributed Cache', `CurrentMetric_DistrCacheUsedConnections` Int64 COMMENT 'Number of currently used connections to Distributed Cache', `CurrentMetric_DistrCacheAllocatedConnections` Int64 COMMENT 'Number of currently allocated connections to Distributed Cache connection pool', `CurrentMetric_DistrCacheBorrowedConnections` Int64 COMMENT 'Number of currently borrowed connections to Distributed Cache connection pool', `CurrentMetric_DistrCacheReadRequests` Int64 COMMENT 'Number of executed Read requests to Distributed Cache', `CurrentMetric_DistrCacheWriteRequests` Int64 COMMENT 'Number of executed Write requests to Distributed Cache', `CurrentMetric_DistrCacheServerConnections` Int64 COMMENT 'Number of open connections to ClickHouse server from Distributed Cache', `CurrentMetric_DistrCacheRegisteredServers` Int64 COMMENT 'Number of distributed cache registered servers', `CurrentMetric_DistrCacheRegisteredServersCurrentAZ` Int64 COMMENT 'Number of distributed cache registered servers in current az', `CurrentMetric_DistrCacheServerS3CachedClients` Int64 COMMENT 'Number of distributed cache S3 cached clients', `CurrentMetric_SchedulerIOReadScheduled` Int64 COMMENT 'Number of IO reads are being scheduled currently', `CurrentMetric_SchedulerIOWriteScheduled` Int64 COMMENT 'Number of IO writes are being scheduled currently', `CurrentMetric_StorageConnectionsStored` Int64 COMMENT 'Total count of sessions stored in the session pool for storages', `CurrentMetric_StorageConnectionsTotal` Int64 COMMENT 'Total count of all sessions: stored in the pool and actively used right now for storages', `CurrentMetric_DiskConnectionsStored` Int64 COMMENT 'Total count of sessions stored in the session pool for disks', `CurrentMetric_DiskConnectionsTotal` Int64 COMMENT 'Total count of all sessions: stored in the pool and actively used right now for disks', `CurrentMetric_HTTPConnectionsStored` Int64 COMMENT 'Total count of sessions stored in the session pool for http hosts', `CurrentMetric_HTTPConnectionsTotal` Int64 COMMENT 'Total count of all sessions: stored in the pool and actively used right now for http hosts', `CurrentMetric_AddressesActive` Int64 COMMENT 'Total count of addresses which are used for creation connections with connection pools', `CurrentMetric_AddressesBanned` Int64 COMMENT 'Total count of addresses which are banned as faulty for creation connections with connection pools', `CurrentMetric_FilteringMarksWithPrimaryKey` Int64 COMMENT 'Number of threads currently doing filtering of mark ranges by the primary key', `CurrentMetric_FilteringMarksWithSecondaryKeys` Int64 COMMENT 'Number of threads currently doing filtering of mark ranges by secondary keys', `CurrentMetric_ConcurrencyControlAcquired` Int64 COMMENT 'Total number of acquired CPU slots', `CurrentMetric_ConcurrencyControlAcquiredNonCompeting` Int64 COMMENT 'Total number of acquired CPU slots that are not considered competing (the first thread if fair_round_robin scheduler is in use)', `CurrentMetric_ConcurrencyControlSoftLimit` Int64 COMMENT 'Value of soft limit on number of CPU slots', `CurrentMetric_DiskS3NoSuchKeyErrors` Int64 COMMENT 'The number of `NoSuchKey` errors that occur when reading data from S3 cloud storage through ClickHouse disks.', `CurrentMetric_SharedCatalogStateApplicationThreads` Int64 COMMENT 'Number of threads in the threadpool for state application in Shared Catalog.', `CurrentMetric_SharedCatalogStateApplicationThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for state application in Shared Catalog.', `CurrentMetric_SharedCatalogStateApplicationThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for state application in Shared Catalog.', `CurrentMetric_SharedCatalogDropLocalThreads` Int64 COMMENT 'Number of threads in the threadpool for drop of local tables in Shared Catalog.', `CurrentMetric_SharedCatalogDropLocalThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for drop of local tables in Shared Catalog.', `CurrentMetric_SharedCatalogDropLocalThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for drop of local tables in Shared Catalog.', `CurrentMetric_SharedCatalogDropZooKeeperThreads` Int64 COMMENT 'Number of threads in the threadpool for drop of object in ZooKeeper in Shared Catalog.', `CurrentMetric_SharedCatalogDropZooKeeperThreadsActive` Int64 COMMENT 'Number of active threads in the threadpool for drop of object in ZooKeeper in Shared Catalog.', `CurrentMetric_SharedCatalogDropZooKeeperThreadsScheduled` Int64 COMMENT 'Number of queued or active jobs in the threadpool for drop of object in ZooKeeper in Shared Catalog.', `CurrentMetric_SharedDatabaseCatalogTablesInLocalDropDetachQueue` Int64 COMMENT 'Number of tables in the queue for local drop or detach in Shared Catalog.', `CurrentMetric_StartupScriptsExecutionState` Int64 COMMENT 'State of startup scripts execution: 0 = not finished, 1 = success, 2 = failure.', `CurrentMetric_IsServerShuttingDown` Int64 COMMENT 'Indicates if the server is shutting down: 0 = no, 1 = yes' ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains history of metrics values from tables system.metrics and system.events, periodically flushed to disk.\n\nIt is safe to truncate or drop this table at any time.' Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default http://localhost:None "POST /v1.46/exec/e36f054f04e41477ceb2978770be6c833abf69df8e988c8e9281ddb82556f016/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/e36f054f04e41477ceb2978770be6c833abf69df8e988c8e9281ddb82556f016/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/e36f054f04e41477ceb2978770be6c833abf69df8e988c8e9281ddb82556f016/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/e36f054f04e41477ceb2978770be6c833abf69df8e988c8e9281ddb82556f016/json HTTP/1.1" 200 586 Executing query SHOW CREATE TABLE system.error_log FORMAT TSVRaw on node_default Executing query SHOW CREATE TABLE system.error_log FORMAT TSVRaw on node_default With storage policy, SHOW CREATE TABLE system.error_log is: CREATE TABLE system.error_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.' CODEC(ZSTD(1)), `event_date` Date COMMENT 'Event date.' CODEC(Delta(2), ZSTD(1)), `event_time` DateTime COMMENT 'Event time.' CODEC(Delta(4), ZSTD(1)), `code` Int32 COMMENT 'Error code.' CODEC(ZSTD(1)), `error` LowCardinality(String) COMMENT 'Error name.' CODEC(ZSTD(1)), `value` UInt64 COMMENT 'Number of errors happened in time interval.' CODEC(ZSTD(3)), `remote` UInt8 COMMENT 'Remote exception (i.e. received during one of the distributed queries).' CODEC(ZSTD(1)) ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains history of error values from table system.errors, periodically flushed to disk.\n\nIt is safe to truncate or drop this table at any time.' With storage policy, SHOW CREATE TABLE system.error_log is: CREATE TABLE system.error_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.' CODEC(ZSTD(1)), `event_date` Date COMMENT 'Event date.' CODEC(Delta(2), ZSTD(1)), `event_time` DateTime COMMENT 'Event time.' CODEC(Delta(4), ZSTD(1)), `code` Int32 COMMENT 'Error code.' CODEC(ZSTD(1)), `error` LowCardinality(String) COMMENT 'Error name.' CODEC(ZSTD(1)), `value` UInt64 COMMENT 'Number of errors happened in time interval.' CODEC(ZSTD(3)), `remote` UInt8 COMMENT 'Remote exception (i.e. received during one of the distributed queries).' CODEC(ZSTD(1)) ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains history of error values from table system.errors, periodically flushed to disk.\n\nIt is safe to truncate or drop this table at any time.' Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query SHOW CREATE TABLE system.latency_log FORMAT TSVRaw on node_default Executing query SHOW CREATE TABLE system.latency_log FORMAT TSVRaw on node_default With storage policy, SHOW CREATE TABLE system.latency_log is: CREATE TABLE system.latency_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `event_date` Date COMMENT 'Event date.', `event_time` DateTime COMMENT 'Event time.', `event_time_microseconds` DateTime64(6) COMMENT 'Event time with microseconds resolution.', `LatencyEvent_S3FirstByteReadAttempt1Microseconds` Array(UInt64) COMMENT 'Time of first byte read from S3 storage (attempt 1).', `LatencyEvent_S3FirstByteWriteAttempt1Microseconds` Array(UInt64) COMMENT 'Time of first byte write to S3 storage (attempt 1).', `LatencyEvent_S3FirstByteReadAttempt2Microseconds` Array(UInt64) COMMENT 'Time of first byte read from S3 storage (attempt 2).', `LatencyEvent_S3FirstByteWriteAttempt2Microseconds` Array(UInt64) COMMENT 'Time of first byte write to S3 storage (attempt 2).', `LatencyEvent_S3FirstByteReadAttemptNMicroseconds` Array(UInt64) COMMENT 'Time of first byte read from S3 storage (attempt N).', `LatencyEvent_S3FirstByteWriteAttemptNMicroseconds` Array(UInt64) COMMENT 'Time of first byte write to S3 storage (attempt N).', `LatencyEvent_S3ConnectMicroseconds` Array(UInt64) COMMENT 'Time to connect for requests to S3 storage.', `LatencyEvent_DiskS3FirstByteReadAttempt1Microseconds` Array(UInt64) COMMENT 'Time of first byte read from DiskS3 storage (attempt 1).', `LatencyEvent_DiskS3FirstByteWriteAttempt1Microseconds` Array(UInt64) COMMENT 'Time of first byte write to DiskS3 storage (attempt 1).', `LatencyEvent_DiskS3FirstByteReadAttempt2Microseconds` Array(UInt64) COMMENT 'Time of first byte read from DiskS3 storage (attempt 2).', `LatencyEvent_DiskS3FirstByteWriteAttempt2Microseconds` Array(UInt64) COMMENT 'Time of first byte write to DiskS3 storage (attempt 2).', `LatencyEvent_DiskS3FirstByteReadAttemptNMicroseconds` Array(UInt64) COMMENT 'Time of first byte read from DiskS3 storage (attempt N).', `LatencyEvent_DiskS3FirstByteWriteAttemptNMicroseconds` Array(UInt64) COMMENT 'Time of first byte write to DiskS3 storage (attempt N).', `LatencyEvent_DiskS3ConnectMicroseconds` Array(UInt64) COMMENT 'Time to connect for requests to DiskS3 storage.' ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains history of all latency buckets, periodically flushed to disk.\n\nIt is safe to truncate or drop this table at any time.' With storage policy, SHOW CREATE TABLE system.latency_log is: CREATE TABLE system.latency_log ( `hostname` LowCardinality(String) COMMENT 'Hostname of the server executing the query.', `event_date` Date COMMENT 'Event date.', `event_time` DateTime COMMENT 'Event time.', `event_time_microseconds` DateTime64(6) COMMENT 'Event time with microseconds resolution.', `LatencyEvent_S3FirstByteReadAttempt1Microseconds` Array(UInt64) COMMENT 'Time of first byte read from S3 storage (attempt 1).', `LatencyEvent_S3FirstByteWriteAttempt1Microseconds` Array(UInt64) COMMENT 'Time of first byte write to S3 storage (attempt 1).', `LatencyEvent_S3FirstByteReadAttempt2Microseconds` Array(UInt64) COMMENT 'Time of first byte read from S3 storage (attempt 2).', `LatencyEvent_S3FirstByteWriteAttempt2Microseconds` Array(UInt64) COMMENT 'Time of first byte write to S3 storage (attempt 2).', `LatencyEvent_S3FirstByteReadAttemptNMicroseconds` Array(UInt64) COMMENT 'Time of first byte read from S3 storage (attempt N).', `LatencyEvent_S3FirstByteWriteAttemptNMicroseconds` Array(UInt64) COMMENT 'Time of first byte write to S3 storage (attempt N).', `LatencyEvent_S3ConnectMicroseconds` Array(UInt64) COMMENT 'Time to connect for requests to S3 storage.', `LatencyEvent_DiskS3FirstByteReadAttempt1Microseconds` Array(UInt64) COMMENT 'Time of first byte read from DiskS3 storage (attempt 1).', `LatencyEvent_DiskS3FirstByteWriteAttempt1Microseconds` Array(UInt64) COMMENT 'Time of first byte write to DiskS3 storage (attempt 1).', `LatencyEvent_DiskS3FirstByteReadAttempt2Microseconds` Array(UInt64) COMMENT 'Time of first byte read from DiskS3 storage (attempt 2).', `LatencyEvent_DiskS3FirstByteWriteAttempt2Microseconds` Array(UInt64) COMMENT 'Time of first byte write to DiskS3 storage (attempt 2).', `LatencyEvent_DiskS3FirstByteReadAttemptNMicroseconds` Array(UInt64) COMMENT 'Time of first byte read from DiskS3 storage (attempt N).', `LatencyEvent_DiskS3FirstByteWriteAttemptNMicroseconds` Array(UInt64) COMMENT 'Time of first byte write to DiskS3 storage (attempt N).', `LatencyEvent_DiskS3ConnectMicroseconds` Array(UInt64) COMMENT 'Time to connect for requests to DiskS3 storage.' ) ENGINE = MergeTree PARTITION BY toYYYYMM(event_date) ORDER BY (event_date, event_time) SETTINGS storage_policy = 'system_tables', index_granularity = 8192 COMMENT 'Contains history of all latency buckets, periodically flushed to disk.\n\nIt is safe to truncate or drop this table at any time.' Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-query_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-query_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-query_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-query_log.xml] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-query_metric_log.xml] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-query_thread_log.xml] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-part_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-part_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-part_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-part_log.xml] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-trace_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-trace_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-trace_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-trace_log.xml] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-metric_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-metric_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-metric_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-metric_log.xml] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-error_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-error_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-error_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-error_log.xml] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-latency_log.xml'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['rm', '/etc/clickhouse-server/config.d/zzz-override-latency_log.xml'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-latency_log.xml] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 rm /etc/clickhouse-server/config.d/zzz-override-latency_log.xml] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 4031 ? 00:00:01 clickhouse Stdout: 4031 ? 00:00:01 clickhouse run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemlogsrecreate-gw9-node_default-1 bash -c pkill clickhouse] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Stdout:4031 Stdout:4031 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Stdout:806 Stdout:806 Clickhouse process running. Clickhouse process running. run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.2.2, port:2181, use_ssl:False get_kazoo_client: zoo2, ip:172.16.2.2, port:2181, use_ssl:False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Stdout:806 Stdout:806 Executing query select 20 on node1 Executing query select 20 on node1 Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Executing query DETACH DATABASE db2 on node1 Executing query DETACH DATABASE db2 on node1 Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-zoo3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.2.3, port:2181, use_ssl:False get_kazoo_client: zoo3, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Executing query DETACH TABLE db3.log on node1 Executing query DETACH TABLE db3.log on node1 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 60 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'Removing access rights for unused directory\''] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 60 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'Removing access rights for unused directory\''] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 60 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'Removing access rights for unused directory'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 60 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'Removing access rights for unused directory'] Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml up -d --no-recreate] Stdout:2025.04.02 04:15:25.815223 [ 825 ] {} DatabaseCatalog: Removing access rights for unused directory store/100/10000000-1000-4000-8000-000000000003/ from disk default (will remove it when timeout exceed) Stdout:2025.04.02 04:15:25.815223 [ 825 ] {} DatabaseCatalog: Removing access rights for unused directory store/100/10000000-1000-4000-8000-000000000003/ from disk default (will remove it when timeout exceed) 1 log line(s) matching "Removing access rights for unused directory" appeared in a 0.069 seconds 1 log line(s) matching "Removing access rights for unused directory" appeared in a 0.069 seconds run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 60 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'directories from store\''] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 60 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'directories from store\''] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 60 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'directories from store'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 60 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'directories from store'] Stdout:2025.04.02 04:15:25.815420 [ 825 ] {} DatabaseCatalog: Cleaned up 3 directories from store/ on disk default Stdout:2025.04.02 04:15:25.815420 [ 825 ] {} DatabaseCatalog: Cleaned up 3 directories from store/ on disk default 1 log line(s) matching "directories from store" appeared in a 0.064 seconds 1 log line(s) matching "directories from store" appeared in a 0.064 seconds run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store] Stdout:0e7 Stdout:0e7 Stdout:100 Stdout:100 Stdout:12 Stdout:12 Stdout:200 Stdout:200 Stdout:300 Stdout:300 Stdout:456 Stdout:456 Stdout:kek Stdout:kek run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store] Stdout:total 24 Stdout:total 24 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 0e7 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 0e7 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 100 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 100 Stdout:---------- 1 root root 0 Apr 2 04:15 12 Stdout:---------- 1 root root 0 Apr 2 04:15 12 Stdout:drwxr-x--- 4 root root 4096 Apr 2 04:15 200 Stdout:drwxr-x--- 4 root root 4096 Apr 2 04:15 200 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 300 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 300 Stdout:drwxr-xr-x 5 root root 4096 Apr 2 04:15 456 Stdout:drwxr-xr-x 5 root root 4096 Apr 2 04:15 456 Stdout:d--------- 2 root root 4096 Apr 2 04:15 kek Stdout:d--------- 2 root root 4096 Apr 2 04:15 kek run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/456'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/456'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/456] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/456] Stdout:total 12 Stdout:total 12 Stdout:d--------- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 Stdout:d--------- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 Stdout:---------- 1 root root 0 Apr 2 04:15 45600000-1000-4000-8000-000000000003 Stdout:---------- 1 root root 0 Apr 2 04:15 45600000-1000-4000-8000-000000000003 Stdout:d--------- 2 root root 4096 Apr 2 04:15 45600000-1000-4000-8000-000000000004 Stdout:d--------- 2 root root 4096 Apr 2 04:15 45600000-1000-4000-8000-000000000004 Stdout:d--------- 2 root root 4096 Apr 2 04:15 testgarbage Stdout:d--------- 2 root root 4096 Apr 2 04:15 testgarbage run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/100'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/100'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/100] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/100] Stdout:10000000-1000-4000-8000-000000000001 Stdout:10000000-1000-4000-8000-000000000001 Stdout:10000000-1000-4000-8000-000000000002 Stdout:10000000-1000-4000-8000-000000000002 Stdout:10000000-1000-4000-8000-000000000003 Stdout:10000000-1000-4000-8000-000000000003 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/100'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/100'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/100] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/100] Stdout:total 12 Stdout:total 12 Stdout:d--------- 2 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000001 Stdout:d--------- 2 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000001 Stdout:d--------- 2 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000002 Stdout:d--------- 2 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000002 Stdout:d--------- 4 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000003 Stdout:d--------- 4 root root 4096 Apr 2 04:15 10000000-1000-4000-8000-000000000003 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/200'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/200'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/200] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/200] Stdout:20000000-1000-4000-8000-000000000001 Stdout:20000000-1000-4000-8000-000000000001 Stdout:20000000-1000-4000-8000-000000000002 Stdout:20000000-1000-4000-8000-000000000002 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/200'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/200'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/200] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/200] Stdout:total 8 Stdout:total 8 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000002 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/300'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/300'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/300] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/300] Stdout:30000000-1000-4000-8000-000000000001 Stdout:30000000-1000-4000-8000-000000000001 Stdout:30000000-1000-4000-8000-000000000002 Stdout:30000000-1000-4000-8000-000000000002 Stdout:30000000-1000-4000-8000-000000000003 Stdout:30000000-1000-4000-8000-000000000003 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/300'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/300'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/300] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/300] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:total 12 Stdout:total 12 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/456'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/456'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/456] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/456] Stdout:4031 Stdout:4031 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Stdout:30000000-1000-4000-8000-000000000003 Stdout:30000000-1000-4000-8000-000000000003 Stdout:45600000-1000-4000-8000-000000000003 Stdout:45600000-1000-4000-8000-000000000003 Stdout:45600000-1000-4000-8000-000000000004 Stdout:45600000-1000-4000-8000-000000000004 Stdout:testgarbage Stdout:testgarbage run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/456'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/456'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/456] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/456] Stdout:total 12 Stdout:total 12 Stdout:d--------- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 Stdout:d--------- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 Stdout:---------- 1 root root 0 Apr 2 04:15 45600000-1000-4000-8000-000000000003 Stdout:---------- 1 root root 0 Apr 2 04:15 45600000-1000-4000-8000-000000000003 Stdout:d--------- 2 root root 4096 Apr 2 04:15 45600000-1000-4000-8000-000000000004 Stdout:d--------- 2 root root 4096 Apr 2 04:15 45600000-1000-4000-8000-000000000004 Stdout:d--------- 2 root root 4096 Apr 2 04:15 testgarbage Stdout:d--------- 2 root root 4096 Apr 2 04:15 testgarbage run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'Removing unused directory\''] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'Removing unused directory\''] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'Removing unused directory'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'Removing unused directory'] Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Running Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Running Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Running Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Running Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Running Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Running Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Creating Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Created Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Starting Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Started Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.6... Waiting for ClickHouse start in node1, ip: 172.16.2.6... http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '0' on node2 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '0' on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node1 Executing query SYSTEM SYNC REPLICA drop_detached_test on node1 http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4c6958b04736663b33c0f0f83856881908a907a997f864d153f9cab0b43a6b7e/json HTTP/1.1" 200 None ClickHouse node1 started ClickHouse node1 started get_instance_ip instance_name=node2 get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.2.5... Waiting for ClickHouse start in node2, ip: 172.16.2.5... http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c2aea9eae0e4fe57e14ba68e20fd0077627d12090dd032d977867e68fee397db/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c2aea9eae0e4fe57e14ba68e20fd0077627d12090dd032d977867e68fee397db/json HTTP/1.1" 200 None ClickHouse node2 started ClickHouse node2 started get_instance_ip instance_name=node3 get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.2.8... Waiting for ClickHouse start in node3, ip: 172.16.2.8... http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/95290de203a1f29eb9652b578fa5dffe1c5587e2d8e9d1eaf9ccc0dd490b30cb/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/95290de203a1f29eb9652b578fa5dffe1c5587e2d8e9d1eaf9ccc0dd490b30cb/json HTTP/1.1" 200 None ClickHouse node3 started ClickHouse node3 started get_instance_ip instance_name=node4 get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node4-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node4 get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node4-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node4, ip: 172.16.2.7... Waiting for ClickHouse start in node4, ip: 172.16.2.7... http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemddlworkerqueue-gw7-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e077b04e9f2a4aa08a695276a122019db899be51f622ef4e6c918a3bb082b9a8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e077b04e9f2a4aa08a695276a122019db899be51f622ef4e6c918a3bb082b9a8/json HTTP/1.1" 200 None ClickHouse node4 started ClickHouse node4 started Executing query CREATE DATABASE testdb on node1 Executing query CREATE DATABASE testdb on node1 Executing query CREATE TABLE testdb.test_table(id UInt32, val String) ENGINE = ReplicatedMergeTree('/clickhouse/test/test_table1', '0') ORDER BY id; on node1 Executing query CREATE TABLE testdb.test_table(id UInt32, val String) ENGINE = ReplicatedMergeTree('/clickhouse/test/test_table1', '0') ORDER BY id; on node1 Executing query CREATE DATABASE testdb on node2 Executing query CREATE DATABASE testdb on node2 Executing query CREATE TABLE testdb.test_table(id UInt32, val String) ENGINE = ReplicatedMergeTree('/clickhouse/test/test_table1', '1') ORDER BY id; on node2 Executing query CREATE TABLE testdb.test_table(id UInt32, val String) ENGINE = ReplicatedMergeTree('/clickhouse/test/test_table1', '1') ORDER BY id; on node2 Executing query CREATE DATABASE testdb on node3 Executing query CREATE DATABASE testdb on node3 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query CREATE TABLE testdb.test_table(id UInt32, val String) ENGINE = ReplicatedMergeTree('/clickhouse/test/test_table2', '0') ORDER BY id; on node3 Executing query CREATE TABLE testdb.test_table(id UInt32, val String) ENGINE = ReplicatedMergeTree('/clickhouse/test/test_table2', '0') ORDER BY id; on node3 Stdout:4031 Stdout:4031 Executing query CREATE DATABASE testdb on node4 Executing query CREATE DATABASE testdb on node4 Executing query CREATE TABLE testdb.test_table(id UInt32, val String) ENGINE = ReplicatedMergeTree('/clickhouse/test/test_table2', '1') ORDER BY id; on node4 Executing query CREATE TABLE testdb.test_table(id UInt32, val String) ENGINE = ReplicatedMergeTree('/clickhouse/test/test_table2', '1') ORDER BY id; on node4 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query INSERT INTO testdb.test_table SELECT number, toString(number) FROM numbers(100) on node1 Executing query INSERT INTO testdb.test_table SELECT number, toString(number) FROM numbers(100) on node1 Executing query INSERT INTO testdb.test_table SELECT number, toString(number) FROM numbers(100) on node3 Executing query INSERT INTO testdb.test_table SELECT number, toString(number) FROM numbers(100) on node3 Executing query SYSTEM SYNC REPLICA testdb.test_table on node2 Executing query SYSTEM SYNC REPLICA testdb.test_table on node2 Executing query SYSTEM SYNC REPLICA testdb.test_table on node4 Executing query SYSTEM SYNC REPLICA testdb.test_table on node4 Executing query ALTER TABLE testdb.test_table ON CLUSTER test_cluster ADD COLUMN somecolumn UInt8 AFTER val on node1 Executing query ALTER TABLE testdb.test_table ON CLUSTER test_cluster ADD COLUMN somecolumn UInt8 AFTER val on node1 Executing query SYSTEM SYNC REPLICA testdb.test_table on node1 Executing query SYSTEM SYNC REPLICA testdb.test_table on node1 Executing query SELECT somecolumn FROM testdb.test_table LIMIT 1 on node1 Executing query SELECT somecolumn FROM testdb.test_table LIMIT 1 on node1 Executing query SELECT If((SELECT count(*) FROM system.distributed_ddl_queue WHERE cluster='test_cluster' AND entry='query-0000000000') > 0, 'ok', 'fail') on node1 Executing query SELECT If((SELECT count(*) FROM system.distributed_ddl_queue WHERE cluster='test_cluster' AND entry='query-0000000000') > 0, 'ok', 'fail') on node1 Executing query SYSTEM SYNC REPLICA testdb.test_table on node2 Executing query SYSTEM SYNC REPLICA testdb.test_table on node2 Executing query SELECT somecolumn FROM testdb.test_table LIMIT 1 on node2 Executing query SELECT somecolumn FROM testdb.test_table LIMIT 1 on node2 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT If((SELECT count(*) FROM system.distributed_ddl_queue WHERE cluster='test_cluster' AND entry='query-0000000000') > 0, 'ok', 'fail') on node2 Executing query SELECT If((SELECT count(*) FROM system.distributed_ddl_queue WHERE cluster='test_cluster' AND entry='query-0000000000') > 0, 'ok', 'fail') on node2 No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemlogsrecreate-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/6d63a1375d50b296537e7c2172451fa068b659126a12cf56f32f66d1ea03816f/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/6d63a1375d50b296537e7c2172451fa068b659126a12cf56f32f66d1ea03816f/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/6d63a1375d50b296537e7c2172451fa068b659126a12cf56f32f66d1ea03816f/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/6d63a1375d50b296537e7c2172451fa068b659126a12cf56f32f66d1ea03816f/json HTTP/1.1" 200 586 Executing query SYSTEM SYNC REPLICA testdb.test_table on node3 Executing query SYSTEM SYNC REPLICA testdb.test_table on node3 Executing query SELECT somecolumn FROM testdb.test_table LIMIT 1 on node3 Executing query SELECT somecolumn FROM testdb.test_table LIMIT 1 on node3 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT If((SELECT count(*) FROM system.distributed_ddl_queue WHERE cluster='test_cluster' AND entry='query-0000000000') > 0, 'ok', 'fail') on node3 Executing query SELECT If((SELECT count(*) FROM system.distributed_ddl_queue WHERE cluster='test_cluster' AND entry='query-0000000000') > 0, 'ok', 'fail') on node3 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system Executing query DROP TABLE IF EXISTS drop_detached_test SYNC on node1 Executing query DROP TABLE IF EXISTS drop_detached_test SYNC on node1 Executing query SYSTEM SYNC REPLICA testdb.test_table on node4 Executing query SYSTEM SYNC REPLICA testdb.test_table on node4 Executing query DROP TABLE IF EXISTS drop_detached_test SYNC on node2 Executing query DROP TABLE IF EXISTS drop_detached_test SYNC on node2 Executing query SELECT somecolumn FROM testdb.test_table LIMIT 1 on node4 Executing query SELECT somecolumn FROM testdb.test_table LIMIT 1 on node4 Executing query CREATE TABLE drop_detached_test ON CLUSTER test_cluster (d UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/drop_detached_test', '{replica}') ORDER BY d PARTITION BY d SETTINGS storage_policy='s3' on node1 Executing query CREATE TABLE drop_detached_test ON CLUSTER test_cluster (d UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/drop_detached_test', '{replica}') ORDER BY d PARTITION BY d SETTINGS storage_policy='s3' on node1 Executing query SELECT If((SELECT count(*) FROM system.distributed_ddl_queue WHERE cluster='test_cluster' AND entry='query-0000000000') > 0, 'ok', 'fail') on node4 Executing query SELECT If((SELECT count(*) FROM system.distributed_ddl_queue WHERE cluster='test_cluster' AND entry='query-0000000000') > 0, 'ok', 'fail') on node4 Executing query INSERT INTO drop_detached_test VALUES (0) on node1 Executing query INSERT INTO drop_detached_test VALUES (0) on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml stop --timeout 20] [gw7] PASSED test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue Executing query ALTER TABLE drop_detached_test FREEZE WITH NAME 'detach_backup1' on node1 Executing query ALTER TABLE drop_detached_test FREEZE WITH NAME 'detach_backup1' on node1 Executing query INSERT INTO drop_detached_test VALUES (1) on node1 Executing query INSERT INTO drop_detached_test VALUES (1) on node1 Executing query ALTER TABLE drop_detached_test FREEZE WITH NAME 'detach_backup2' on node1 Executing query ALTER TABLE drop_detached_test FREEZE WITH NAME 'detach_backup2' on node1 run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Stdout:4873 Stdout:4873 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogsrecreate-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogsrecreate-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:4873 Stdout:4873 Executing query select 20 on node_default Executing query select 20 on node_default Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/.env --project-name rootteststorageurlwithproxy-gw8 --file /ClickHouse/tests/integration/test_storage_url_with_proxy/_instances-0-gw8/proxy_list_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/detach_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/detach_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/detach_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Executing query SYSTEM FLUSH LOGS on node_default Executing query SYSTEM FLUSH LOGS on node_default run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/detach_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/detach_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/detach_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Executing query SYSTEM UNFREEZE WITH NAME 'detach_backup2' on node1 Executing query SYSTEM UNFREEZE WITH NAME 'detach_backup2' on node1 Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SHOW CREATE TABLE system.query_log on node_default Executing query SYSTEM UNFREEZE WITH NAME 'detach_backup1' on node1 Executing query SYSTEM UNFREEZE WITH NAME 'detach_backup1' on node1 Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query ALTER TABLE drop_detached_test DETACH PARTITION '0' on node1 Executing query ALTER TABLE drop_detached_test DETACH PARTITION '0' on node1 Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query SHOW CREATE TABLE system.query_metric_log on node_default Executing query ALTER TABLE drop_detached_test DETACH PARTITION '1' on node1 Executing query ALTER TABLE drop_detached_test DETACH PARTITION '1' on node1 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-resolver-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-proxy_list_node-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-minio1-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Stopping Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Stopped Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Removing Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-proxy2-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Removed Stderr: Container rootteststorageurlwithproxy-gw8-proxy1-1 Removed Stderr: Volume rootteststorageurlwithproxy-gw8_data1-1 Removing Stderr: Volume rootteststorageurlwithproxy-gw8_data1-1 Removing Stderr: Network rootteststorageurlwithproxy-gw8_default Removing Stderr: Network rootteststorageurlwithproxy-gw8_default Removing Stderr: Volume rootteststorageurlwithproxy-gw8_data1-1 Removed Stderr: Volume rootteststorageurlwithproxy-gw8_data1-1 Removed Stderr: Network rootteststorageurlwithproxy-gw8_default Removed Stderr: Network rootteststorageurlwithproxy-gw8_default Removed Cleanup called Cleanup called Executing query SHOW CREATE TABLE system.query_thread_log on node_default Executing query SHOW CREATE TABLE system.query_thread_log on node_default Docker networks for project rootteststorageurlwithproxy-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageurlwithproxy-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststorageurlwithproxy-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageurlwithproxy-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststorageurlwithproxy-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststorageurlwithproxy-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageurlwithproxy-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageurlwithproxy-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageurlwithproxy-gw8 No running containers for project: rootteststorageurlwithproxy-gw8 Trying to prune unused networks... Trying to prune unused networks... Executing query SHOW CREATE TABLE system.query_thread_log on node_default Executing query SHOW CREATE TABLE system.query_thread_log on node_default Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 test_structured_logging_json/test.py::test_structured_logging_json_format Running tests in /ClickHouse/tests/integration/test_structured_logging_json/test.py Running tests in /ClickHouse/tests/integration/test_structured_logging_json/test.py Cluster start called. is_up=False Cluster start called. is_up=False Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default Docker networks for project rootteststructuredloggingjson-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststructuredloggingjson-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststructuredloggingjson-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststructuredloggingjson-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststructuredloggingjson-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststructuredloggingjson-gw8 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project rootteststructuredloggingjson-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststructuredloggingjson-gw8 are NETWORK ID NAME DRIVER SCOPE Executing query SHOW CREATE TABLE system.part_log on node_default Executing query SHOW CREATE TABLE system.part_log on node_default Docker containers for project rootteststructuredloggingjson-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststructuredloggingjson-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststructuredloggingjson-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststructuredloggingjson-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststructuredloggingjson-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststructuredloggingjson-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststructuredloggingjson-gw8 No running containers for project: rootteststructuredloggingjson-gw8 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Executing query SHOW CREATE TABLE system.part_log on node_default Executing query SHOW CREATE TABLE system.part_log on node_default Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 Setup directory for instance: node_all_keys Setup directory for instance: node_all_keys Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_structured_logging_json/configs/config_all_keys_json.xml'] to /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_structured_logging_json/configs/config_all_keys_json.xml'] to /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/configs/config.d Setup database dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/database Setup database dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/database Setup logs dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/logs Setup logs dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node_some_keys Setup directory for instance: node_some_keys Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_structured_logging_json/configs/config_some_keys_json.xml'] to /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_structured_logging_json/configs/config_some_keys_json.xml'] to /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/configs/config.d Setup database dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/database Setup database dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/database Setup logs dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/logs Setup logs dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node_no_keys Setup directory for instance: node_no_keys Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_structured_logging_json/configs/config_no_keys_json.xml'] to /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_structured_logging_json/configs/config_no_keys_json.xml'] to /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/configs/config.d Setup database dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/database Setup database dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/database Setup logs dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/logs Setup logs dir /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default Command:[docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml pull] Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SHOW CREATE TABLE system.trace_log on node_default Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Executing query SHOW CREATE TABLE system.metric_log on node_default Executing query SHOW CREATE TABLE system.metric_log on node_default Executing query SHOW CREATE TABLE system.metric_log on node_default Executing query SHOW CREATE TABLE system.metric_log on node_default Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Executing query SHOW CREATE TABLE system.error_log on node_default Executing query SHOW CREATE TABLE system.error_log on node_default Executing query SHOW CREATE TABLE system.error_log on node_default Executing query SHOW CREATE TABLE system.error_log on node_default Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW CREATE TABLE system.latency_log on node_default Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default Executing query SYSTEM FLUSH LOGS on node_default Executing query SYSTEM FLUSH LOGS on node_default Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'query_thread_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'part_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'trace_log%' on node_default Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Connected to Mongo dbs: ['admin', 'config', 'local'] Connected to Mongo dbs: ['admin', 'config', 'local'] ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml up -d --no-recreate] Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '1' on node2 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '1' on node2 Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'metric_log%' on node_default Executing query SYSTEM SYNC REPLICA drop_detached_test on node1 Executing query SYSTEM SYNC REPLICA drop_detached_test on node1 Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'error_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default Executing query SHOW TABLES FROM system LIKE 'latency_log%' on node_default Executing query DROP TABLE IF EXISTS system.query_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.query_log_0 sync on node_default Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Creating Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Running Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Created Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Starting Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Started Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.9.5... Waiting for ClickHouse start in node, ip: 172.16.9.5... http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttablefunctionmongodblegacy-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS system.query_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.query_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.query_log_2 sync on node_default Executing query DROP TABLE IF EXISTS system.query_log_2 sync on node_default http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS system.query_metric_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.query_metric_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.query_metric_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.query_metric_log_1 sync on node_default http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24d2167e451458ce5e0638c9d591fd7b3bfb9f0f30d129fe2960a405b52d5681/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query DROP TABLE IF EXISTS system.query_metric_log_2 sync on node_default Executing query DROP TABLE IF EXISTS system.query_metric_log_2 sync on node_default Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (0, '0x0') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (0, '0x0') on node Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/.env --project-name roottestsystemddlworkerqueue-gw7 --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_system_ddl_worker_queue/_instances-0-gw7/node4/docker-compose.yml down --volumes] Executing query DROP TABLE IF EXISTS system.query_thread_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.query_thread_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.query_thread_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.query_thread_log_1 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (1, '0x1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (1, '0x1') on node Executing query DROP TABLE IF EXISTS system.query_thread_log_2 sync on node_default Executing query DROP TABLE IF EXISTS system.query_thread_log_2 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (2, '0x4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (2, '0x4') on node Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query DROP TABLE IF EXISTS system.part_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.part_log_0 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (3, '0x9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (3, '0x9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (4, '0x10') on node Executing query DROP TABLE IF EXISTS system.part_log_1 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (4, '0x10') on node Executing query DROP TABLE IF EXISTS system.part_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.part_log_2 sync on node_default Executing query DROP TABLE IF EXISTS system.part_log_2 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (5, '0x19') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (5, '0x19') on node Executing query DROP TABLE IF EXISTS system.trace_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.trace_log_0 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (6, '0x24') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (6, '0x24') on node Executing query DROP TABLE IF EXISTS system.trace_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.trace_log_1 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (7, '0x31') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (7, '0x31') on node Executing query DROP TABLE IF EXISTS system.trace_log_2 sync on node_default Executing query DROP TABLE IF EXISTS system.trace_log_2 sync on node_default Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-node1-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-node4-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-node2-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-node3-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Stopping Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Stopped Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Removing Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-zoo3-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-zoo1-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Removed Stderr: Container roottestsystemddlworkerqueue-gw7-zoo2-1 Removed Stderr: Network roottestsystemddlworkerqueue-gw7_default Removing Stderr: Network roottestsystemddlworkerqueue-gw7_default Removing Stderr: Network roottestsystemddlworkerqueue-gw7_default Removed Stderr: Network roottestsystemddlworkerqueue-gw7_default Removed Cleanup called Cleanup called Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (8, '0x40') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (8, '0x40') on node Docker networks for project roottestsystemddlworkerqueue-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemddlworkerqueue-gw7 are NETWORK ID NAME DRIVER SCOPE Executing query DROP TABLE IF EXISTS system.metric_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.metric_log_0 sync on node_default Docker containers for project roottestsystemddlworkerqueue-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemddlworkerqueue-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemddlworkerqueue-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemddlworkerqueue-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemddlworkerqueue-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemddlworkerqueue-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemddlworkerqueue-gw7 No running containers for project: roottestsystemddlworkerqueue-gw7 Trying to prune unused networks... Trying to prune unused networks... Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (9, '0x51') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (9, '0x51') on node Executing query DROP TABLE IF EXISTS system.metric_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.metric_log_1 sync on node_default Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (10, '0x64') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (10, '0x64') on node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated Running tests in /ClickHouse/tests/integration/test_system_logs_hostname/test_replicated.py Running tests in /ClickHouse/tests/integration/test_system_logs_hostname/test_replicated.py Cluster start called. is_up=False Cluster start called. is_up=False Executing query DROP TABLE IF EXISTS system.metric_log_2 sync on node_default Executing query DROP TABLE IF EXISTS system.metric_log_2 sync on node_default Docker networks for project roottestsystemlogshostnamereplicated-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogshostnamereplicated-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemlogshostnamereplicated-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogshostnamereplicated-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemlogshostnamereplicated-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogshostnamereplicated-gw7 are DRIVER VOLUME NAME Cleanup called Cleanup called Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (11, '0x79') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (11, '0x79') on node Docker networks for project roottestsystemlogshostnamereplicated-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogshostnamereplicated-gw7 are NETWORK ID NAME DRIVER SCOPE Executing query DROP TABLE IF EXISTS system.error_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.error_log_0 sync on node_default Docker containers for project roottestsystemlogshostnamereplicated-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogshostnamereplicated-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemlogshostnamereplicated-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogshostnamereplicated-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemlogshostnamereplicated-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemlogshostnamereplicated-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemlogshostnamereplicated-gw7 No running containers for project: roottestsystemlogshostnamereplicated-gw7 Trying to prune unused networks... Trying to prune unused networks... Executing query DROP TABLE IF EXISTS system.error_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.error_log_1 sync on node_default Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (12, '0x90') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (12, '0x90') on node Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 Setup directory for instance: node1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_logs_hostname/configs/replicated_servers.xml'] to /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_logs_hostname/configs/replicated_servers.xml'] to /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/database Setup database dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/database Setup logs dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/logs Setup logs dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_logs_hostname/configs/replicated_servers.xml'] to /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_logs_hostname/configs/replicated_servers.xml'] to /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/configs/config.d Executing query DROP TABLE IF EXISTS system.error_log_2 sync on node_default Executing query DROP TABLE IF EXISTS system.error_log_2 sync on node_default Setup database dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/database Setup database dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/database Setup logs dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/logs Setup logs dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node3 Setup directory for instance: node3 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_logs_hostname/configs/replicated_servers.xml'] to /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_logs_hostname/configs/replicated_servers.xml'] to /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/database Setup database dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/database Setup logs dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/logs Setup logs dir /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml pull] Executing query DROP TABLE IF EXISTS system.latency_log_0 sync on node_default Executing query DROP TABLE IF EXISTS system.latency_log_0 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (13, '0xa9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (13, '0xa9') on node Executing query DROP TABLE IF EXISTS system.latency_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.latency_log_1 sync on node_default Executing query DROP TABLE IF EXISTS system.latency_log_2 sync on node_default Executing query DROP TABLE IF EXISTS system.latency_log_2 sync on node_default Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (14, '0xc4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (14, '0xc4') on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml stop --timeout 20] [gw9] PASSED test_system_logs_recreate/test.py::test_system_logs_recreate Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (15, '0xe1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (15, '0xe1') on node Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (16, '0x100') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (16, '0x100') on node Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '1' on node1 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '1' on node1 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (17, '0x121') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (17, '0x121') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (18, '0x144') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (18, '0x144') on node Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Stopping Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Stopping Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Stopped Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/.env --project-name roottestsystemlogsrecreate-gw9 --file /ClickHouse/tests/integration/test_system_logs_recreate/_instances-0-gw9/node_default/docker-compose.yml down --volumes] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (19, '0x169') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (19, '0x169') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (20, '0x190') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (20, '0x190') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (21, '0x1b9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (21, '0x1b9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (22, '0x1e4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (22, '0x1e4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (23, '0x211') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (23, '0x211') on node Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Stopping Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Stopping Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Stopped Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Stopped Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Removing Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Removing Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Removed Stderr: Container roottestsystemlogsrecreate-gw9-node_default-1 Removed Stderr: Network roottestsystemlogsrecreate-gw9_default Removing Stderr: Network roottestsystemlogsrecreate-gw9_default Removing Stderr: Network roottestsystemlogsrecreate-gw9_default Removed Stderr: Network roottestsystemlogsrecreate-gw9_default Removed Cleanup called Cleanup called Docker networks for project roottestsystemlogsrecreate-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogsrecreate-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemlogsrecreate-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogsrecreate-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (24, '0x240') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (24, '0x240') on node Docker volumes for project roottestsystemlogsrecreate-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogsrecreate-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemlogsrecreate-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemlogsrecreate-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemlogsrecreate-gw9 No running containers for project: roottestsystemlogsrecreate-gw9 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (25, '0x271') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (25, '0x271') on node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 test_system_logs_comment/test.py::test_system_logs_comment Running tests in /ClickHouse/tests/integration/test_system_logs_comment/test.py Running tests in /ClickHouse/tests/integration/test_system_logs_comment/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottestsystemlogscomment-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogscomment-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemlogscomment-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogscomment-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemlogscomment-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogscomment-gw9 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottestsystemlogscomment-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogscomment-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemlogscomment-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogscomment-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (26, '0x2a4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (26, '0x2a4') on node Docker volumes for project roottestsystemlogscomment-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogscomment-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemlogscomment-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemlogscomment-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemlogscomment-gw9 No running containers for project: roottestsystemlogscomment-gw9 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:13 Stdout:13 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (27, '0x2d9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (27, '0x2d9') on node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 13 Volumes pruned: 13 Setup directory for instance: node_default Setup directory for instance: node_default Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/configs/config.d Copy custom test config files [] to /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/database Setup database dir /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/database Setup logs dir /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/logs Setup logs dir /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml pull] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (28, '0x310') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (28, '0x310') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (29, '0x349') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (29, '0x349') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (30, '0x384') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (30, '0x384') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (31, '0x3c1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (31, '0x3c1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (32, '0x400') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (32, '0x400') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (33, '0x441') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (33, '0x441') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (34, '0x484') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (34, '0x484') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (35, '0x4c9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (35, '0x4c9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (36, '0x510') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (36, '0x510') on node Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/.env --project-name rootteststorageazureblobstoragecluster-gw2 --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_1/docker-compose.yml --file /ClickHouse/tests/integration/test_storage_azure_blob_storage/_instances-cluster-0-gw2/node_2/docker-compose.yml down --volumes] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (37, '0x559') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (37, '0x559') on node Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '0' on node1 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '0' on node1 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (38, '0x5a4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (38, '0x5a4') on node Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query SYSTEM SYNC REPLICA drop_detached_test on node2 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (39, '0x5f1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (39, '0x5f1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (40, '0x640') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (40, '0x640') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (41, '0x691') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (41, '0x691') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (42, '0x6e4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (42, '0x6e4') on node Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Removing Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Removing Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Removing Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Removing Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Removing Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Removing Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Removed Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_1-1 Removed Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Removed Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_0-1 Removed Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Removed Stderr: Container rootteststorageazureblobstoragecluster-gw2-node_2-1 Removed Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Stopping Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Stopped Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Removing Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Removing Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Removed Stderr: Container rootteststorageazureblobstoragecluster-gw2-azurite1-1 Removed Stderr: Volume rootteststorageazureblobstoragecluster-gw2_data1-1 Removing Stderr: Volume rootteststorageazureblobstoragecluster-gw2_data1-1 Removing Stderr: Network rootteststorageazureblobstoragecluster-gw2_default Removing Stderr: Network rootteststorageazureblobstoragecluster-gw2_default Removing Stderr: Volume rootteststorageazureblobstoragecluster-gw2_data1-1 Removed Stderr: Volume rootteststorageazureblobstoragecluster-gw2_data1-1 Removed Stderr: Network rootteststorageazureblobstoragecluster-gw2_default Removed Stderr: Network rootteststorageazureblobstoragecluster-gw2_default Removed Cleanup called Cleanup called Docker networks for project rootteststorageazureblobstoragecluster-gw2 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorageazureblobstoragecluster-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststorageazureblobstoragecluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorageazureblobstoragecluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (43, '0x739') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (43, '0x739') on node Docker volumes for project rootteststorageazureblobstoragecluster-gw2 are DRIVER VOLUME NAME Docker volumes for project rootteststorageazureblobstoragecluster-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorageazureblobstoragecluster-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorageazureblobstoragecluster-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorageazureblobstoragecluster-gw2 No running containers for project: rootteststorageazureblobstoragecluster-gw2 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:11 Stdout:11 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (44, '0x790') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (44, '0x790') on node Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 11 Volumes pruned: 11 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (45, '0x7e9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (45, '0x7e9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (46, '0x844') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (46, '0x844') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (47, '0x8a1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (47, '0x8a1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (48, '0x900') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (48, '0x900') on node Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (49, '0x961') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (49, '0x961') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (50, '0x9c4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (50, '0x9c4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (51, '0xa29') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (51, '0xa29') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (52, '0xa90') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (52, '0xa90') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (53, '0xaf9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (53, '0xaf9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (54, '0xb64') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (54, '0xb64') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (55, '0xbd1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (55, '0xbd1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (56, '0xc40') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (56, '0xc40') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (57, '0xcb1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (57, '0xcb1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (58, '0xd24') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (58, '0xd24') on node Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '0' on node2 Executing query ALTER TABLE drop_detached_test DROP DETACHED PARTITION '0' on node2 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (59, '0xd99') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (59, '0xd99') on node Executing query SYSTEM SYNC REPLICA drop_detached_test on node1 Executing query SYSTEM SYNC REPLICA drop_detached_test on node1 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (60, '0xe10') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (60, '0xe10') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (61, '0xe89') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (61, '0xe89') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (62, '0xf04') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (62, '0xf04') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (63, '0xf81') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (63, '0xf81') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (64, '0x1000') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (64, '0x1000') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (65, '0x1081') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (65, '0x1081') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (66, '0x1104') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (66, '0x1104') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (67, '0x1189') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (67, '0x1189') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (68, '0x1210') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (68, '0x1210') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (69, '0x1299') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (69, '0x1299') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (70, '0x1324') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (70, '0x1324') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (71, '0x13b1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (71, '0x13b1') on node Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node1 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (72, '0x1440') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (72, '0x1440') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (73, '0x14d1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (73, '0x14d1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (74, '0x1564') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (74, '0x1564') on node Stderr: node_all_keys Skipped - Image is already being pulled by node_no_keys Stderr: node_all_keys Skipped - Image is already being pulled by node_no_keys Stderr: node_some_keys Skipped - Image is already being pulled by node_no_keys Stderr: node_some_keys Skipped - Image is already being pulled by node_no_keys Stderr: node_no_keys Pulling Stderr: node_no_keys Pulling Stderr: node_no_keys Pulled Stderr: node_no_keys Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml up -d --no-recreate] Stderr: node_default Pulling Stderr: node_default Pulling Stderr: node_default Pulled Stderr: node_default Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml up -d --no-recreate] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (75, '0x15f9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (75, '0x15f9') on node Stderr: node3 Skipped - Image is already being pulled by node2 Stderr: node3 Skipped - Image is already being pulled by node2 Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: zoo3 Skipped - Image is already being pulled by node2 Stderr: zoo3 Skipped - Image is already being pulled by node2 Stderr: zoo1 Skipped - Image is already being pulled by node2 Stderr: zoo1 Skipped - Image is already being pulled by node2 Stderr: zoo2 Skipped - Image is already being pulled by node2 Stderr: zoo2 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulling Stderr: node2 Pulled Stderr: node2 Pulled Setup ZooKeeper Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/log', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/config', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/coordination', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/log', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/config', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/coordination', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/log', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/config', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/coordination'] Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/log', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/config', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper1/coordination', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/log', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/config', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper2/coordination', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/log', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/config', '/ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/keeper3/coordination'] Command:[docker compose --project-name roottestsystemlogshostnamereplicated-gw7 --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Command:[docker compose --project-name roottestsystemlogshostnamereplicated-gw7 --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (76, '0x1690') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (76, '0x1690') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (77, '0x1729') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (77, '0x1729') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (78, '0x17c4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (78, '0x17c4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (79, '0x1861') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (79, '0x1861') on node Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='drop_detached_test' AND is_done=0 on node2 Stderr: Network rootteststructuredloggingjson-gw8_default Creating Stderr: Network rootteststructuredloggingjson-gw8_default Creating Stderr: Network rootteststructuredloggingjson-gw8_default Created Stderr: Network rootteststructuredloggingjson-gw8_default Created Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Creating Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Creating Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Creating Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Creating Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Creating Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Creating Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Created Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Created Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Created Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Created Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Created Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Created Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Starting Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Starting Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Starting Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Starting Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Starting Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Starting Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Started Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Started Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Started Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Started Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Started Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node_all_keys get_instance_ip instance_name=node_all_keys http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_all_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_all_keys-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_all_keys get_instance_ip instance_name=node_all_keys http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_all_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_all_keys-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_all_keys, ip: 172.16.1.4... Waiting for ClickHouse start in node_all_keys, ip: 172.16.1.4... http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_all_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_all_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (80, '0x1900') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (80, '0x1900') on node [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation Executing query DROP TABLE IF EXISTS zero_copy_mutation SYNC on node1 Executing query DROP TABLE IF EXISTS zero_copy_mutation SYNC on node1 Stderr: Network roottestsystemlogscomment-gw9_default Creating Stderr: Network roottestsystemlogscomment-gw9_default Creating Stderr: Network roottestsystemlogscomment-gw9_default Created Stderr: Network roottestsystemlogscomment-gw9_default Created Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Creating Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Creating Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Created Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Created Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Starting Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Starting Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Started Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node_default get_instance_ip instance_name=node_default http://localhost:None "GET /v1.46/containers/roottestsystemlogscomment-gw9-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogscomment-gw9-node_default-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_default get_instance_ip instance_name=node_default http://localhost:None "GET /v1.46/containers/roottestsystemlogscomment-gw9-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogscomment-gw9-node_default-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_default, ip: 172.16.2.2... Waiting for ClickHouse start in node_default, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottestsystemlogscomment-gw9-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogscomment-gw9-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS zero_copy_mutation SYNC on node2 Executing query DROP TABLE IF EXISTS zero_copy_mutation SYNC on node2 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (81, '0x19a1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (81, '0x19a1') on node http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None Executing query CREATE TABLE zero_copy_mutation (id UInt64, value1 String, value2 String, value3 String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/zero_copy_mutation', '{replica}') ORDER BY id PARTITION BY (id % 4) SETTINGS storage_policy='s3', old_parts_lifetime=1000 on node1 Executing query CREATE TABLE zero_copy_mutation (id UInt64, value1 String, value2 String, value3 String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/zero_copy_mutation', '{replica}') ORDER BY id PARTITION BY (id % 4) SETTINGS storage_policy='s3', old_parts_lifetime=1000 on node1 http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None Executing query CREATE TABLE zero_copy_mutation (id UInt64, value1 String, value2 String, value3 String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/zero_copy_mutation', '{replica}') ORDER BY id PARTITION BY (id % 4) SETTINGS storage_policy='s3', old_parts_lifetime=1000 on node2 Executing query CREATE TABLE zero_copy_mutation (id UInt64, value1 String, value2 String, value3 String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/zero_copy_mutation', '{replica}') ORDER BY id PARTITION BY (id % 4) SETTINGS storage_policy='s3', old_parts_lifetime=1000 on node2 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (82, '0x1a44') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (82, '0x1a44') on node Stderr: Container roottests3tablefunctions-gw5-node-1 Stopping Stderr: Container roottests3tablefunctions-gw5-node-1 Stopping Stderr: Container roottests3tablefunctions-gw5-resolver-1 Stopping Stderr: Container roottests3tablefunctions-gw5-resolver-1 Stopping Stderr: Container roottests3tablefunctions-gw5-node-1 Stopped Stderr: Container roottests3tablefunctions-gw5-node-1 Stopped Stderr: Container roottests3tablefunctions-gw5-minio1-1 Stopping Stderr: Container roottests3tablefunctions-gw5-minio1-1 Stopping Stderr: Container roottests3tablefunctions-gw5-minio1-1 Stopped Stderr: Container roottests3tablefunctions-gw5-minio1-1 Stopped Stderr: Container roottests3tablefunctions-gw5-resolver-1 Stopped Stderr: Container roottests3tablefunctions-gw5-resolver-1 Stopped Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Stopping Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Stopping Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Stopping Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Stopping Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Stopped Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Stopped Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Stopped Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (83, '0x1ae9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (83, '0x1ae9') on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/.env --project-name roottests3tablefunctions-gw5 --file /ClickHouse/tests/integration/test_s3_table_functions/_instances-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a82bdae04c433520781fcd04584880fd2b263450129dc4b154d87c4083c4ec3d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None ClickHouse node_all_keys started ClickHouse node_all_keys started get_instance_ip instance_name=node_some_keys get_instance_ip instance_name=node_some_keys http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_some_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_some_keys-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_some_keys get_instance_ip instance_name=node_some_keys http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_some_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_some_keys-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_some_keys, ip: 172.16.1.3... Waiting for ClickHouse start in node_some_keys, ip: 172.16.1.3... http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_some_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_some_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8c00bd9fe3c6eec303ed2b46dec5644ef8e1ffa308683dddb36dbb252a0660f4/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8c00bd9fe3c6eec303ed2b46dec5644ef8e1ffa308683dddb36dbb252a0660f4/json HTTP/1.1" 200 None ClickHouse node_some_keys started ClickHouse node_some_keys started get_instance_ip instance_name=node_no_keys get_instance_ip instance_name=node_no_keys http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_no_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_no_keys-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_no_keys get_instance_ip instance_name=node_no_keys http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_no_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_no_keys-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_no_keys, ip: 172.16.1.2... Waiting for ClickHouse start in node_no_keys, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_no_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/rootteststructuredloggingjson-gw8-node_no_keys-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2c2a1438c69b05866514b3f197725358d9b311a60c800d2b409bc97aa79dfea0/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2c2a1438c69b05866514b3f197725358d9b311a60c800d2b409bc97aa79dfea0/json HTTP/1.1" 200 None ClickHouse node_no_keys started ClickHouse node_no_keys started run container_id:rootteststructuredloggingjson-gw8-node_all_keys-1 detach:False nothrow:False cmd: ['cat', '/etc/clickhouse-server/config.d/config_all_keys_json.xml'] run container_id:rootteststructuredloggingjson-gw8-node_all_keys-1 detach:False nothrow:False cmd: ['cat', '/etc/clickhouse-server/config.d/config_all_keys_json.xml'] Command:[docker exec rootteststructuredloggingjson-gw8-node_all_keys-1 cat /etc/clickhouse-server/config.d/config_all_keys_json.xml] Command:[docker exec rootteststructuredloggingjson-gw8-node_all_keys-1 cat /etc/clickhouse-server/config.d/config_all_keys_json.xml] Executing query INSERT INTO zero_copy_mutation SELECT * FROM generateRandom('id UInt64, value1 String, value2 String, value3 String') limit 1000000 on node1 Executing query INSERT INTO zero_copy_mutation SELECT * FROM generateRandom('id UInt64, value1 String, value2 String, value3 String') limit 1000000 on node1 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: --> Stdout: Stdout: Stdout: json Stdout: json Stdout: Stdout: Stdout: DATE_TIME Stdout: DATE_TIME Stdout: DATE_TIME_UTC Stdout: DATE_TIME_UTC Stdout: THREAD_NAME Stdout: THREAD_NAME Stdout: THREAD_ID Stdout: THREAD_ID Stdout: LEVEL Stdout: LEVEL Stdout: QUERY_ID Stdout: QUERY_ID Stdout: LOGGER_NAME Stdout: LOGGER_NAME Stdout: MESSAGE Stdout: MESSAGE Stdout: SOURCE_FILE Stdout: SOURCE_FILE Stdout: SOURCE_LINE Stdout: SOURCE_LINE Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: run container_id:rootteststructuredloggingjson-gw8-node_some_keys-1 detach:False nothrow:False cmd: ['cat', '/etc/clickhouse-server/config.d/config_some_keys_json.xml'] run container_id:rootteststructuredloggingjson-gw8-node_some_keys-1 detach:False nothrow:False cmd: ['cat', '/etc/clickhouse-server/config.d/config_some_keys_json.xml'] Command:[docker exec rootteststructuredloggingjson-gw8-node_some_keys-1 cat /etc/clickhouse-server/config.d/config_some_keys_json.xml] Command:[docker exec rootteststructuredloggingjson-gw8-node_some_keys-1 cat /etc/clickhouse-server/config.d/config_some_keys_json.xml] http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41a972ee25aaba63ec7b84435c55e7366d9d432a79f0a805fec40492d9784e4c/json HTTP/1.1" 200 None ClickHouse node_default started ClickHouse node_default started run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = MergeTree\n PARTITION BY (event_date)\n ORDER BY (event_time)\n TTL event_date + INTERVAL 14 DAY DELETE\n SETTINGS ttl_only_drop_parts=1\n COMMENT \'test_comment\'\n \n \n \n \n " > /etc/clickhouse-server/config.d/yyy-override-query_log.xml\n '] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo "\n \n \n ENGINE = MergeTree\n PARTITION BY (event_date)\n ORDER BY (event_time)\n TTL event_date + INTERVAL 14 DAY DELETE\n SETTINGS ttl_only_drop_parts=1\n COMMENT \'test_comment\'\n \n \n \n \n " > /etc/clickhouse-server/config.d/yyy-override-query_log.xml\n '] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c echo " ENGINE = MergeTree PARTITION BY (event_date) ORDER BY (event_time) TTL event_date + INTERVAL 14 DAY DELETE SETTINGS ttl_only_drop_parts=1 COMMENT 'test_comment' " > /etc/clickhouse-server/config.d/yyy-override-query_log.xml ] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c echo " ENGINE = MergeTree PARTITION BY (event_date) ORDER BY (event_time) TTL event_date + INTERVAL 14 DAY DELETE SETTINGS ttl_only_drop_parts=1 COMMENT 'test_comment' " > /etc/clickhouse-server/config.d/yyy-override-query_log.xml ] Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: --> Stdout: Stdout: Stdout: json Stdout: json Stdout: Stdout: Stdout: DATE_TIME Stdout: DATE_TIME Stdout: DATE_TIME_UTC Stdout: DATE_TIME_UTC Stdout: THREAD_NAME Stdout: THREAD_NAME Stdout: THREAD_ID Stdout: THREAD_ID Stdout: LEVEL Stdout: LEVEL Stdout: Stdout: LOGGER_NAME--> Stdout: MESSAGE Stdout: MESSAGE Stdout: SOURCE_FILE Stdout: SOURCE_FILE Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: run container_id:rootteststructuredloggingjson-gw8-node_no_keys-1 detach:False nothrow:False cmd: ['cat', '/etc/clickhouse-server/config.d/config_no_keys_json.xml'] run container_id:rootteststructuredloggingjson-gw8-node_no_keys-1 detach:False nothrow:False cmd: ['cat', '/etc/clickhouse-server/config.d/config_no_keys_json.xml'] Command:[docker exec rootteststructuredloggingjson-gw8-node_no_keys-1 cat /etc/clickhouse-server/config.d/config_no_keys_json.xml] Command:[docker exec rootteststructuredloggingjson-gw8-node_no_keys-1 cat /etc/clickhouse-server/config.d/config_no_keys_json.xml] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestsystemlogscomment-gw9-node_default-1 bash -c ps -C clickhouse] Command:[docker exec -u root roottestsystemlogscomment-gw9-node_default-1 bash -c ps -C clickhouse] Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: --> Stdout: Stdout: Stdout: json Stdout: json Stdout: Stdout: --> Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Executing query SELECT 1 on node_all_keys Executing query SELECT 1 on node_all_keys Stdout: PID TTY TIME CMD Stdout: PID TTY TIME CMD Stdout: 9 ? 00:00:00 clickhouse Stdout: 9 ? 00:00:00 clickhouse run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestsystemlogscomment-gw9-node_default-1 bash -c pkill clickhouse] Command:[docker exec -u root roottestsystemlogscomment-gw9-node_default-1 bash -c pkill clickhouse] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:9 Stdout:9 Stderr:time="2025-04-02T04:15:42Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:15:42Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestsystemlogshostnamereplicated-gw7_default Creating Stderr: Network roottestsystemlogshostnamereplicated-gw7_default Creating Stderr: Network roottestsystemlogshostnamereplicated-gw7_default Created Stderr: Network roottestsystemlogshostnamereplicated-gw7_default Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Started Stderr:time="2025-04-02T04:15:43Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:43Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:43Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:15:43Z" level=debug msg="otel error" error="" Wait ZooKeeper to start Wait ZooKeeper to start get_instance_ip instance_name=zoo1 get_instance_ip instance_name=zoo1 grep in log called grep in log called run container_id:rootteststructuredloggingjson-gw8-node_all_keys-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true'] run container_id:rootteststructuredloggingjson-gw8-node_all_keys-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true'] Command:[docker exec rootteststructuredloggingjson-gw8-node_all_keys-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true] Command:[docker exec rootteststructuredloggingjson-gw8-node_all_keys-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true] http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-zoo1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.3.4, port:2181, use_ssl:False get_kazoo_client: zoo1, ip:172.16.3.4, port:2181, use_ssl:False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (84, '0x1b90') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (84, '0x1b90') on node Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.815901","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"SentryWriter","MESSAGE":"Sending crash reports is disabled","SOURCE_FILE":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)","SOURCE_LINE":"144"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.815901","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"SentryWriter","MESSAGE":"Sending crash reports is disabled","SOURCE_FILE":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)","SOURCE_LINE":"144"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","SOURCE_FILE":"","SOURCE_LINE":"0"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","SOURCE_FILE":"","SOURCE_LINE":"0"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873632","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"starting up","SOURCE_FILE":"","SOURCE_LINE":"0"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873632","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"starting up","SOURCE_FILE":"","SOURCE_LINE":"0"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873641","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","SOURCE_FILE":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)","SOURCE_LINE":"585"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873641","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","SOURCE_FILE":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)","SOURCE_LINE":"585"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873735","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Jemalloc","MESSAGE":"Value for background_thread set to true (from true)","SOURCE_FILE":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]","SOURCE_LINE":"32"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873735","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Jemalloc","MESSAGE":"Value for background_thread set to true (from true)","SOURCE_FILE":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]","SOURCE_LINE":"32"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877381","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1027"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877381","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1027"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877418","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1042"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877418","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1042"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877455","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Pipe","MESSAGE":"Pipe capacity is 1.00 MiB","SOURCE_FILE":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)","SOURCE_LINE":"131"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877455","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Pipe","MESSAGE":"Pipe capacity is 1.00 MiB","SOURCE_FILE":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)","SOURCE_LINE":"131"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878183","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsReader","MESSAGE":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)","SOURCE_LINE":"214"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878183","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsReader","MESSAGE":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)","SOURCE_LINE":"214"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878344","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/class\/thermal","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()","SOURCE_LINE":"126"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878344","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/class\/thermal","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()","SOURCE_LINE":"126"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878367","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/block","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()","SOURCE_LINE":"163"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878367","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/block","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()","SOURCE_LINE":"163"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878479","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/devices\/system\/edac","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()","SOURCE_LINE":"190"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878479","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/devices\/system\/edac","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()","SOURCE_LINE":"190"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878503","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/class\/hwmon","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()","SOURCE_LINE":"220"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878503","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/class\/hwmon","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()","SOURCE_LINE":"220"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.028504","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1349"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.028504","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1349"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.028629","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1403"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.028629","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1403"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032548","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1407"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032548","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1407"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032616","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"MemoryWorker","MESSAGE":"Starting background memory thread with period of 50ms, using Cgroups as source","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()","SOURCE_LINE":"252"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032616","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"MemoryWorker","MESSAGE":"Starting background memory thread with period of 50ms, using Cgroups as source","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()","SOURCE_LINE":"252"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032706","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgSchPool","MESSAGE":"Create BackgroundSchedulePool with 512 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032706","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgSchPool","MESSAGE":"Create BackgroundSchedulePool with 512 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.086483","THREAD_NAME":"","THREAD_ID":"14","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"MemoryTracker","MESSAGE":"Correcting the value of global memory tracker from 3.53 MiB to 116.19 MiB","SOURCE_FILE":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)","SOURCE_LINE":"542"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.086483","THREAD_NAME":"","THREAD_ID":"14","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"MemoryTracker","MESSAGE":"Correcting the value of global memory tracker from 3.53 MiB to 116.19 MiB","SOURCE_FILE":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)","SOURCE_LINE":"542"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087254","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"rlimit on number of file descriptors is 1048576","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1474"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087254","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"rlimit on number of file descriptors is 1048576","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1474"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087291","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"rlimit on number of threads is 18446744073709551615","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1496"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087291","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"rlimit on number of threads is 18446744073709551615","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1496"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087316","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initializing DateLUT.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1524"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087316","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initializing DateLUT.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1524"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initialized DateLUT with time zone 'Etc\/UTC'.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1526"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initialized DateLUT with time zone 'Etc\/UTC'.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1526"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087375","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)","SOURCE_LINE":"1324"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087375","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)","SOURCE_LINE":"1324"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087281","THREAD_NAME":"","THREAD_ID":"17","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"CancellationChecker","MESSAGE":"Started worker function","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()","SOURCE_LINE":"97"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087281","THREAD_NAME":"","THREAD_ID":"17","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"CancellationChecker","MESSAGE":"Started worker function","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()","SOURCE_LINE":"97"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087814","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_all_keys' as replica host.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1605"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087814","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_all_keys' as replica host.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1605"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087848","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initializing interserver credentials.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1619"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087848","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initializing interserver credentials.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1619"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087982","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"NamedCollectionsMetadataStorage","MESSAGE":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)","SOURCE_LINE":"627"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087982","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"NamedCollectionsMetadataStorage","MESSAGE":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)","SOURCE_LINE":"627"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088013","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)","SOURCE_LINE":"256"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088013","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)","SOURCE_LINE":"256"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088026","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from sql","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)","SOURCE_LINE":"276"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088026","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from sql","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)","SOURCE_LINE":"276"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088043","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"FileCacheFactory","MESSAGE":"Will load 0 caches from default cache config","SOURCE_FILE":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)","SOURCE_LINE":"226"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088043","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"FileCacheFactory","MESSAGE":"Will load 0 caches from default cache config","SOURCE_FILE":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)","SOURCE_LINE":"226"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.090409","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loading config '\/etc\/clickhouse-server\/config.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"128"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.090409","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loading config '\/etc\/clickhouse-server\/config.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"128"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.090431","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"680"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.090431","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"680"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091652","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091652","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091787","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091787","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091895","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091895","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092137","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_all_keys_json.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092137","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_all_keys_json.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.099047","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","SOURCE_LINE":"926"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.099047","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","SOURCE_LINE":"926"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.099173","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"175"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.099173","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"175"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.107990","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1792"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.107990","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1792"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1819"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1819"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Merges and mutations memory limit is set to 15.30 GiB","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1832"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Merges and mutations memory limit is set to 15.30 GiB","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1832"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108202","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting max_remote_read_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1874"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108202","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting max_remote_read_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1874"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting max_remote_write_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1875"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting max_remote_write_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1875"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108230","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1886"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108230","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1886"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108250","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgBufSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108250","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgBufSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110088","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgMBSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110088","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgMBSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.111654","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgDistSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.111654","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgDistSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.114578","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)","SOURCE_LINE":"267"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.114578","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)","SOURCE_LINE":"267"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.115649","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"193"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.115649","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"193"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.115700","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","SOURCE_LINE":"45"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.115700","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","SOURCE_LINE":"45"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.116188","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2187"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.116188","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2187"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118842","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loading config '\/etc\/clickhouse-server\/users.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"128"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118842","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loading config '\/etc\/clickhouse-server\/users.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"128"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118861","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"680"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118861","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"680"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119082","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119082","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119124","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119124","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119674","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","SOURCE_LINE":"926"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119674","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","SOURCE_LINE":"926"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119708","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"175"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119708","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"175"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120173","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"193"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120173","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"193"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120462","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","SOURCE_LINE":"45"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120462","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","SOURCE_LINE":"45"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120872","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Access(user directories)","MESSAGE":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)","SOURCE_LINE":"359"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120872","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Access(user directories)","MESSAGE":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)","SOURCE_LINE":"359"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123056","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Access(local_directory)","MESSAGE":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()","SOURCE_LINE":"246"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123056","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Access(local_directory)","MESSAGE":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()","SOURCE_LINE":"246"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123081","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Access(local_directory)","MESSAGE":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)","SOURCE_LINE":"186"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123081","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Access(local_directory)","MESSAGE":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)","SOURCE_LINE":"186"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123239","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Access(user directories)","MESSAGE":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)","SOURCE_LINE":"397"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123239","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Access(user directories)","MESSAGE":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)","SOURCE_LINE":"397"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsMemoryUsageObserver","MESSAGE":"Started cgroup current memory usage observer thread","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()","SOURCE_LINE":"43"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsMemoryUsageObserver","MESSAGE":"Started cgroup current memory usage observer thread","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()","SOURCE_LINE":"43"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123724","THREAD_NAME":"","THREAD_ID":"582","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsMemoryUsageObserver","MESSAGE":"Memory amount initially available to the process is 30.60 GiB","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()","SOURCE_LINE":"67"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123724","THREAD_NAME":"","THREAD_ID":"582","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsMemoryUsageObserver","MESSAGE":"Memory amount initially available to the process is 30.60 GiB","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()","SOURCE_LINE":"67"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.125466","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5940"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.125466","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5940"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.126540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for move operations with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5950"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.126540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for move operations with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5950"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.127948","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for fetches with num_threads=16, num_tasks=16","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5960"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.127948","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for fetches with num_threads=16, num_tasks=16","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5960"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.128664","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5970"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.128664","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5970"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","SOURCE_LINE":"932"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","SOURCE_LINE":"932"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130702","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","SOURCE_LINE":"932"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130702","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","SOURCE_LINE":"932"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130937","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DNSCacheUpdater","MESSAGE":"Update period 15 seconds","SOURCE_FILE":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()","SOURCE_LINE":"49"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130937","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DNSCacheUpdater","MESSAGE":"Update period 15 seconds","SOURCE_FILE":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()","SOURCE_LINE":"49"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130959","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Loading metadata from \/var\/lib\/clickhouse\/","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2315"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130959","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Loading metadata from \/var\/lib\/clickhouse\/","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2315"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131002","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Database disk name: default","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","SOURCE_LINE":"1182"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131002","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Database disk name: default","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","SOURCE_LINE":"1182"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131009","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Database disk name: default, path: \/var\/lib\/clickhouse\/","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","SOURCE_LINE":"1190"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131009","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Database disk name: default, path: \/var\/lib\/clickhouse\/","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","SOURCE_LINE":"1190"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131043","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/system.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131043","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/system.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131002","THREAD_NAME":"","THREAD_ID":"15","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DNSResolver","MESSAGE":"Updating DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","SOURCE_LINE":"446"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131002","THREAD_NAME":"","THREAD_ID":"15","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DNSResolver","MESSAGE":"Updating DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","SOURCE_LINE":"446"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131092","THREAD_NAME":"","THREAD_ID":"15","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DNSResolver","MESSAGE":"Updated DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","SOURCE_LINE":"477"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131092","THREAD_NAME":"","THREAD_ID":"15","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DNSResolver","MESSAGE":"Updated DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","SOURCE_LINE":"477"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.132288","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/information_schema.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.132288","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/information_schema.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.133960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.133960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135316","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","SOURCE_LINE":"311"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135316","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","SOURCE_LINE":"311"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TablesLoader","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 7.2185e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","SOURCE_LINE":"63"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TablesLoader","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 7.2185e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","SOURCE_LINE":"63"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ReferentialDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ReferentialDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135353","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"LoadingDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135353","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"LoadingDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135410","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","SOURCE_LINE":"693"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135410","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","SOURCE_LINE":"693"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135419","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","SOURCE_LINE":"693"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135419","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","SOURCE_LINE":"693"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135461","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Atomic database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135461","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Atomic database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135470","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: none -> 0","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135470","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: none -> 0","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135479","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #1 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135479","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #1 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135618","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Wait load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)","SOURCE_LINE":"802"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135618","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Wait load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)","SOURCE_LINE":"802"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135672","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Ordinary database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135672","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Ordinary database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135725","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Ordinary database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135725","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Ordinary database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135736","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #2 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135736","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #2 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135803","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135803","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135866","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Atomic database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135866","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Atomic database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135880","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135880","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135915","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135915","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135961","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: 0 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135961","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: 0 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.139132","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_log from query_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.139132","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_log from query_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.140221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_thread_log from query_thread_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.140221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_thread_log from query_thread_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.140682","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.part_log from part_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.140682","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.part_log from part_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141008","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.trace_log from trace_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141008","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.trace_log from trace_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141281","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.crash_log from crash_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141281","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.crash_log from crash_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141448","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.text_log from text_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141448","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.text_log from text_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141703","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.metric_log from metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141703","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.metric_log from metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146046","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.latency_log from latency_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146046","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.latency_log from latency_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146341","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.error_log from error_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146341","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.error_log from error_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146584","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146584","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146610","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146610","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146627","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.s3queue_log from s3queue_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146627","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.s3queue_log from s3queue_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146814","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146814","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146830","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.asynchronous_metric_log from asynchronous_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146830","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.asynchronous_metric_log from asynchronous_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147010","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.opentelemetry_span_log from opentelemetry_span_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147010","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.opentelemetry_span_log from opentelemetry_span_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_views_log from query_views_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_views_log from query_views_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147574","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147574","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147603","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.session_log since corresponding section 'session_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147603","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.session_log since corresponding section 'session_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147615","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147615","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]","SOURCE_LINE":"153"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147626","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.processors_profile_log from processors_profile_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147626","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.processors_profile_log from processors_profile_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.asynchronous_insert_log from asynchronous_insert_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.asynchronous_insert_log from asynchronous_insert_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.backup_log from backup_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.backup_log from backup_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148572","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.blob_storage_log from blob_storage_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148572","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.blob_storage_log from blob_storage_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148807","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_metric_log from query_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148807","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_metric_log from query_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]","SOURCE_LINE":"158"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.153508","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.153508","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.153540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.153540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154821","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154821","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154834","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154834","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154855","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154855","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154860","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154860","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154881","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154881","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154887","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154887","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154906","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154906","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154911","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154911","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154949","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154949","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154962","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154962","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155060","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155060","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155067","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155067","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155112","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155112","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155118","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155118","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155170","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155170","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155216","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155216","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155222","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155222","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155254","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155254","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155260","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155260","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155343","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155343","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155379","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155379","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155384","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155384","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155421","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155421","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155428","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155428","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155547","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155547","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155588","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155588","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155594","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155594","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155619","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155619","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155629","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155629","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155660","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155660","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155683","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155683","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155688","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155688","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155708","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155708","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155733","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155733","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155738","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155738","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155760","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155760","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155768","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155768","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155795","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155795","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155830","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155830","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155854","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155854","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155860","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155860","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155922","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155922","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155928","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155928","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155955","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155955","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156141","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156141","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156152","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156152","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156181","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156181","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156187","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156187","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156209","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156209","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156213","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156213","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156237","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156237","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156242","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156242","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156271","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156271","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156276","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156276","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156348","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156348","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156357","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156357","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156395","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156395","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156400","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156400","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156484","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156484","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156493","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156493","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156544","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156544","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156550","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156550","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156630","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156630","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156636","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156636","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156712","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156712","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156723","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156723","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156762","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156762","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156767","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156767","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156891","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156891","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156898","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156898","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156928","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156928","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156934","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156934","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156954","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156954","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156958","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156958","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156984","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156984","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156994","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156994","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157028","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157028","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157056","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157056","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157061","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157061","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157080","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157080","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157085","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157085","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157135","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157135","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157186","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157186","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157219","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157219","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157417","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157417","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157431","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157431","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157534","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157534","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157542","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157542","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157582","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157582","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157587","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157587","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157607","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157607","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157614","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157614","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157659","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157659","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157665","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157665","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157707","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157707","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157737","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157737","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157742","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157742","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157761","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157761","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157766","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157766","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157914","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157914","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158062","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158062","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158075","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158075","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158141","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158141","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158263","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158263","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158270","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158270","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158412","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158412","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158430","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158430","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158508","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158508","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158516","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158516","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158555","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158555","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158561","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158561","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158706","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158706","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158716","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158716","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158776","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158776","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158782","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158782","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158827","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158827","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158833","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158833","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158868","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158868","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158873","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158873","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158912","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158912","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158918","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158918","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159051","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159051","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159061","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159061","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159117","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159117","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159123","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159123","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159172","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159172","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159222","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159222","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159283","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159283","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159289","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159289","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159399","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159399","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159406","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159406","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159501","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159501","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159510","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159510","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159546","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159546","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159551","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159551","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159582","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159582","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159587","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159587","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159624","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159624","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159633","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159633","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159669","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159669","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159674","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159674","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159716","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159716","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159780","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159780","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159786","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159786","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159831","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159831","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159858","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159858","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159863","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159863","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159905","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159905","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159910","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159910","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159942","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159942","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159947","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159947","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160002","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160002","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160008","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160008","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160081","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160081","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160088","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160088","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160116","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160116","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160121","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160121","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160157","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160157","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160190","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160190","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160196","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160196","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160230","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160230","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160235","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160235","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160261","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160261","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160266","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160266","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160288","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160288","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160293","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160293","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160354","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160354","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160365","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160365","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160420","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160420","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160437","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160437","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160467","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160467","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160472","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160472","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160549","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160549","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160556","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160556","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.167597","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseCatalog","MESSAGE":"Found 0 partially dropped tables. Will load them and retry removal.","SOURCE_FILE":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()","SOURCE_LINE":"1050"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.167597","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseCatalog","MESSAGE":"Found 0 partially dropped tables. Will load them and retry removal.","SOURCE_FILE":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()","SOURCE_LINE":"1050"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168775","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (default)","MESSAGE":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","SOURCE_LINE":"311"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168775","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (default)","MESSAGE":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","SOURCE_LINE":"311"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168788","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TablesLoader","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 2.616e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","SOURCE_LINE":"63"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168788","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TablesLoader","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 2.616e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","SOURCE_LINE":"63"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168800","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ReferentialDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168800","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ReferentialDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"LoadingDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"LoadingDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168832","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"loadMetadata","MESSAGE":"Start asynchronous loading of databases","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)","SOURCE_LINE":"268"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168832","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"loadMetadata","MESSAGE":"Start asynchronous loading of databases","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)","SOURCE_LINE":"268"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168841","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168841","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168846","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Atomic database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168846","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Atomic database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168851","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: none -> 2","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168851","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: none -> 2","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168856","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #1 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168856","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #1 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168975","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"UserDefinedSQLObjectsLoaderFromDisk","MESSAGE":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","SOURCE_LINE":"131"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168975","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"UserDefinedSQLObjectsLoaderFromDisk","MESSAGE":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","SOURCE_LINE":"131"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168990","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"UserDefinedSQLObjectsLoaderFromDisk","MESSAGE":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","SOURCE_LINE":"135"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168990","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"UserDefinedSQLObjectsLoaderFromDisk","MESSAGE":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","SOURCE_LINE":"135"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"WorkloadEntityDiskStorage","MESSAGE":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","SOURCE_LINE":"129"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"WorkloadEntityDiskStorage","MESSAGE":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","SOURCE_LINE":"129"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169031","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"WorkloadEntityDiskStorage","MESSAGE":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","SOURCE_LINE":"133"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169031","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"WorkloadEntityDiskStorage","MESSAGE":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","SOURCE_LINE":"133"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169046","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Loaded metadata.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2408"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169046","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Loaded metadata.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2408"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169040","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Ordinary database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169040","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Ordinary database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169095","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Ordinary database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169095","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Ordinary database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169105","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #2 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169105","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #2 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169137","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"bool DB::(anonymous namespace)::checkPermissionsImpl()","MESSAGE":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()","SOURCE_LINE":"216"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169137","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"bool DB::(anonymous namespace)::checkPermissionsImpl()","MESSAGE":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()","SOURCE_LINE":"216"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169157","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Tasks stats provider: procfs","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2428"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169157","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Tasks stats provider: procfs","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2428"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169179","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Atomic database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169179","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Atomic database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169190","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Atomic database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169190","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Atomic database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169197","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169197","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169204","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169204","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169220","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169220","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169283","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169283","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169359","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: 2 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169359","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: 2 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169474","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","SOURCE_LINE":"37"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169474","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","SOURCE_LINE":"37"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169517","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","SOURCE_LINE":"48"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169517","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","SOURCE_LINE":"48"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169524","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Generating new RSA key pair.","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()","SOURCE_LINE":"107"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169524","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Generating new RSA key pair.","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()","SOURCE_LINE":"107"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.182770","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CertificateReloader","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","SOURCE_LINE":"142"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.182770","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CertificateReloader","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","SOURCE_LINE":"142"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.182790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CertificateReloader","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","SOURCE_LINE":"142"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.182790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CertificateReloader","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","SOURCE_LINE":"142"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185601","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for http:\/\/0.0.0.0:8123","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185601","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for http:\/\/0.0.0.0:8123","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185651","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for native protocol (tcp): 0.0.0.0:9000","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185651","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for native protocol (tcp): 0.0.0.0:9000","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185697","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185697","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185746","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185746","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185753","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Ready for connections.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2548"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185753","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Ready for connections.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2548"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.221518","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"Poco","MESSAGE":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","SOURCE_LINE":"49"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.221518","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"Poco","MESSAGE":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","SOURCE_LINE":"49"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223495","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCPHandlerFactory","MESSAGE":"TCP Request. Address: 172.16.1.1:57458","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","SOURCE_LINE":"54"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223495","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCPHandlerFactory","MESSAGE":"TCP Request. Address: 172.16.1.1:57458","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","SOURCE_LINE":"54"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223585","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Client has not sent any data.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","SOURCE_LINE":"330"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223585","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Client has not sent any data.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","SOURCE_LINE":"330"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223592","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","SOURCE_LINE":"2630"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223592","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","SOURCE_LINE":"2630"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.449993","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"Poco","MESSAGE":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","SOURCE_LINE":"49"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.449993","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"Poco","MESSAGE":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","SOURCE_LINE":"49"} Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (85, '0x1c39') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (85, '0x1c39') on node Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450067","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCPHandlerFactory","MESSAGE":"TCP Request. Address: 172.16.1.1:57466","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","SOURCE_LINE":"54"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450067","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCPHandlerFactory","MESSAGE":"TCP Request. Address: 172.16.1.1:57466","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","SOURCE_LINE":"54"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450147","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()","SOURCE_LINE":"1695"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450147","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()","SOURCE_LINE":"1695"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450170","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"Authenticating user 'default' from 172.16.1.1:57466","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","SOURCE_LINE":"364"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450170","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"Authenticating user 'default' from 172.16.1.1:57466","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","SOURCE_LINE":"364"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450219","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","SOURCE_LINE":"374"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450219","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","SOURCE_LINE":"374"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450234","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()","SOURCE_LINE":"534"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450234","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()","SOURCE_LINE":"534"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450356","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"432"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450356","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"432"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450388","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"433"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450388","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"433"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450415","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"434"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450415","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"434"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.463932","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const","SOURCE_LINE":"664"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.463932","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const","SOURCE_LINE":"664"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464321","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"executeQuery","MESSAGE":"(from 172.16.1.1:57466) (query 1, line 1) SELECT 1 (stage: Complete)","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)","SOURCE_LINE":"237"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464321","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"executeQuery","MESSAGE":"(from 172.16.1.1:57466) (query 1, line 1) SELECT 1 (stage: Complete)","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)","SOURCE_LINE":"237"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464367","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"CancellationChecker","MESSAGE":"Did not add the task because the timeout is 0. Query: SELECT 1","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)","SOURCE_LINE":"77"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464367","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"CancellationChecker","MESSAGE":"Did not add the task because the timeout is 0. Query: SELECT 1","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)","SOURCE_LINE":"77"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464563","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"Planner","MESSAGE":"Query to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()","SOURCE_LINE":"1296"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464563","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"Planner","MESSAGE":"Query to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()","SOURCE_LINE":"1296"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464667","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"Planner","MESSAGE":"Query from stage FetchColumns to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()","SOURCE_LINE":"1574"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464667","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"Planner","MESSAGE":"Query from stage FetchColumns to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()","SOURCE_LINE":"1574"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464817","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"QueryMetricLog","MESSAGE":"Scheduling next collecting task for query_id fce58dd8-584b-422e-b892-32498fffbd41 in 999 ms","SOURCE_FILE":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)","SOURCE_LINE":"219"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464817","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"QueryMetricLog","MESSAGE":"Scheduling next collecting task for query_id fce58dd8-584b-422e-b892-32498fffbd41 in 999 ms","SOURCE_FILE":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)","SOURCE_LINE":"219"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.465688","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"executeQuery","MESSAGE":"Read 1 rows, 1.00 B in 0.001369 sec., 730.4601899196493 rows\/sec., 730.46 B\/sec.","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)","SOURCE_LINE":"592"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.465688","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"executeQuery","MESSAGE":"Read 1 rows, 1.00 B in 0.001369 sec., 730.4601899196493 rows\/sec., 730.46 B\/sec.","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)","SOURCE_LINE":"592"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.465806","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"TCPHandler","MESSAGE":"Processed in 0.002027772 sec.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)","SOURCE_LINE":"865"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.465806","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"TCPHandler","MESSAGE":"Processed in 0.002027772 sec.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)","SOURCE_LINE":"865"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481884","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Closing connection (open: true, cancelled: false, eof: true)","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","SOURCE_LINE":"456"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481884","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Closing connection (open: true, cancelled: false, eof: true)","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","SOURCE_LINE":"456"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481920","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","SOURCE_LINE":"2630"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481920","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","SOURCE_LINE":"2630"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481958","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; DB::Session::~Session()","SOURCE_LINE":"308"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481958","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; DB::Session::~Session()","SOURCE_LINE":"308"} grep result {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.815901","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"SentryWriter","MESSAGE":"Sending crash reports is disabled","SOURCE_FILE":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)","SOURCE_LINE":"144"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","SOURCE_FILE":"","SOURCE_LINE":"0"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873632","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"starting up","SOURCE_FILE":"","SOURCE_LINE":"0"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873641","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","SOURCE_FILE":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)","SOURCE_LINE":"585"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873735","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Jemalloc","MESSAGE":"Value for background_thread set to true (from true)","SOURCE_FILE":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]","SOURCE_LINE":"32"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877381","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1027"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877418","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1042"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877455","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Pipe","MESSAGE":"Pipe capacity is 1.00 MiB","SOURCE_FILE":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)","SOURCE_LINE":"131"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878183","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsReader","MESSAGE":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)","SOURCE_LINE":"214"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878344","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/class\/thermal","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()","SOURCE_LINE":"126"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878367","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/block","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()","SOURCE_LINE":"163"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878479","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/devices\/system\/edac","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()","SOURCE_LINE":"190"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878503","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/class\/hwmon","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()","SOURCE_LINE":"220"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.028504","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1349"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.028629","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1403"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032548","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1407"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032616","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"MemoryWorker","MESSAGE":"Starting background memory thread with period of 50ms, using Cgroups as source","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()","SOURCE_LINE":"252"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032706","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgSchPool","MESSAGE":"Create BackgroundSchedulePool with 512 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.086483","THREAD_NAME":"","THREAD_ID":"14","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"MemoryTracker","MESSAGE":"Correcting the value of global memory tracker from 3.53 MiB to 116.19 MiB","SOURCE_FILE":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)","SOURCE_LINE":"542"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087254","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"rlimit on number of file descriptors is 1048576","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1474"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087291","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"rlimit on number of threads is 18446744073709551615","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1496"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087316","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initializing DateLUT.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1524"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initialized DateLUT with time zone 'Etc\/UTC'.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1526"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087375","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)","SOURCE_LINE":"1324"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087281","THREAD_NAME":"","THREAD_ID":"17","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"CancellationChecker","MESSAGE":"Started worker function","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()","SOURCE_LINE":"97"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087814","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_all_keys' as replica host.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1605"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087848","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initializing interserver credentials.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1619"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087982","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"NamedCollectionsMetadataStorage","MESSAGE":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)","SOURCE_LINE":"627"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088013","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)","SOURCE_LINE":"256"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088026","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from sql","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)","SOURCE_LINE":"276"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088043","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"FileCacheFactory","MESSAGE":"Will load 0 caches from default cache config","SOURCE_FILE":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)","SOURCE_LINE":"226"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.090409","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loading config '\/etc\/clickhouse-server\/config.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"128"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.090431","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"680"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091652","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091787","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091895","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092137","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_all_keys_json.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.099047","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","SOURCE_LINE":"926"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.099173","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"175"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.107990","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1792"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1819"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Merges and mutations memory limit is set to 15.30 GiB","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1832"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108202","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting max_remote_read_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1874"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting max_remote_write_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1875"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108230","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1886"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108250","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgBufSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110088","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgMBSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.111654","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgDistSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.114578","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)","SOURCE_LINE":"267"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.115649","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"193"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.115700","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","SOURCE_LINE":"45"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.116188","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2187"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118842","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loading config '\/etc\/clickhouse-server\/users.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"128"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118861","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"680"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119082","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119124","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119674","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","SOURCE_LINE":"926"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119708","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"175"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120173","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"193"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120462","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","SOURCE_LINE":"45"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120872","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Access(user directories)","MESSAGE":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)","SOURCE_LINE":"359"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123056","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Access(local_directory)","MESSAGE":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()","SOURCE_LINE":"246"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123081","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Access(local_directory)","MESSAGE":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)","SOURCE_LINE":"186"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123239","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Access(user directories)","MESSAGE":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)","SOURCE_LINE":"397"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsMemoryUsageObserver","MESSAGE":"Started cgroup current memory usage observer thread","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()","SOURCE_LINE":"43"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123724","THREAD_NAME":"","THREAD_ID":"582","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsMemoryUsageObserver","MESSAGE":"Memory amount initially available to the process is 30.60 GiB","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()","SOURCE_LINE":"67"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.125466","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5940"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.126540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for move operations with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5950"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.127948","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for fetches with num_threads=16, num_tasks=16","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5960"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.128664","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5970"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","SOURCE_LINE":"932"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130702","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","SOURCE_LINE":"932"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130937","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DNSCacheUpdater","MESSAGE":"Update period 15 seconds","SOURCE_FILE":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()","SOURCE_LINE":"49"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130959","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Loading metadata from \/var\/lib\/clickhouse\/","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2315"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131002","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Database disk name: default","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","SOURCE_LINE":"1182"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131009","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Database disk name: default, path: \/var\/lib\/clickhouse\/","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","SOURCE_LINE":"1190"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131043","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/system.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131002","THREAD_NAME":"","THREAD_ID":"15","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DNSResolver","MESSAGE":"Updating DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","SOURCE_LINE":"446"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131092","THREAD_NAME":"","THREAD_ID":"15","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DNSResolver","MESSAGE":"Updated DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","SOURCE_LINE":"477"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.132288","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/information_schema.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.133960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135316","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","SOURCE_LINE":"311"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TablesLoader","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 7.2185e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","SOURCE_LINE":"63"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ReferentialDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135353","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"LoadingDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135410","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","SOURCE_LINE":"693"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135419","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","SOURCE_LINE":"693"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135461","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Atomic database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135470","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: none -> 0","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135479","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #1 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135618","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Wait load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)","SOURCE_LINE":"802"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135672","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Ordinary database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135725","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Ordinary database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135736","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #2 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135803","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135866","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Atomic database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135880","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135915","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135961","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: 0 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.139132","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_log from query_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.140221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_thread_log from query_thread_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.140682","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.part_log from part_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141008","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.trace_log from trace_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141281","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.crash_log from crash_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141448","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.text_log from text_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141703","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.metric_log from metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146046","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.latency_log from latency_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146341","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.error_log from error_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146584","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146610","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146627","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.s3queue_log from s3queue_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146814","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146830","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.asynchronous_metric_log from asynchronous_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147010","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.opentelemetry_span_log from opentelemetry_span_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_views_log from query_views_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147574","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147603","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.session_log since corresponding section 'session_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147615","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147626","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.processors_profile_log from processors_profile_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.asynchronous_insert_log from asynchronous_insert_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.backup_log from backup_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148572","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.blob_storage_log from blob_storage_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148807","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_metric_log from query_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.153508","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.153540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154821","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154834","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154855","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154860","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154881","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154887","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154906","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154911","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154949","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154962","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155060","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155067","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155112","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155118","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155170","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155216","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155222","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155254","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155260","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155343","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155379","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155384","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155421","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155428","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155547","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155588","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155594","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155619","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155629","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155660","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155683","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155688","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155708","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155733","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155738","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155760","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155768","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155795","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155830","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155854","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155860","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155922","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155928","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155955","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156141","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156152","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156181","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156187","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156209","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156213","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156237","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156242","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156271","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156276","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156348","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156357","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156395","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156400","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156484","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156493","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156544","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156550","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156630","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156636","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156712","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156723","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156762","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156767","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156891","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156898","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156928","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156934","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156954","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156958","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156984","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156994","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157028","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157056","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157061","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157080","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157085","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157135","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157186","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157219","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157417","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157431","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157534","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157542","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157582","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157587","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157607","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157614","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157659","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157665","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157707","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157737","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157742","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157761","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157766","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157914","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158062","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158075","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158141","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158263","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158270","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158412","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158430","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158508","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158516","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158555","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158561","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158706","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158716","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158776","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158782","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158827","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158833","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158868","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158873","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158912","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158918","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159051","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159061","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159117","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159123","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159172","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159222","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159283","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159289","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159399","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159406","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159501","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159510","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159546","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159551","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159582","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159587","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159624","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159633","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159669","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159674","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159716","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159780","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159786","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159831","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159858","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159863","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159905","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159910","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159942","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159947","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160002","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160008","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160081","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160088","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160116","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160121","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160157","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160190","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160196","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160230","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160235","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160261","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160266","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160288","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160293","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160354","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160365","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160420","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160437","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160467","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160472","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160549","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160556","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.167597","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseCatalog","MESSAGE":"Found 0 partially dropped tables. Will load them and retry removal.","SOURCE_FILE":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()","SOURCE_LINE":"1050"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168775","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (default)","MESSAGE":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","SOURCE_LINE":"311"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168788","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TablesLoader","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 2.616e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","SOURCE_LINE":"63"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168800","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ReferentialDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"LoadingDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168832","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"loadMetadata","MESSAGE":"Start asynchronous loading of databases","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)","SOURCE_LINE":"268"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168841","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168846","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Atomic database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168851","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: none -> 2","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168856","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #1 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168975","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"UserDefinedSQLObjectsLoaderFromDisk","MESSAGE":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","SOURCE_LINE":"131"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168990","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"UserDefinedSQLObjectsLoaderFromDisk","MESSAGE":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","SOURCE_LINE":"135"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"WorkloadEntityDiskStorage","MESSAGE":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","SOURCE_LINE":"129"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169031","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"WorkloadEntityDiskStorage","MESSAGE":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","SOURCE_LINE":"133"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169046","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Loaded metadata.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2408"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169040","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Ordinary database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169095","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Ordinary database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169105","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #2 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169137","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"bool DB::(anonymous namespace)::checkPermissionsImpl()","MESSAGE":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()","SOURCE_LINE":"216"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169157","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Tasks stats provider: procfs","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2428"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169179","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Atomic database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169190","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Atomic database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169197","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169204","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169220","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169283","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169359","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: 2 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169474","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","SOURCE_LINE":"37"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169517","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","SOURCE_LINE":"48"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169524","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Generating new RSA key pair.","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()","SOURCE_LINE":"107"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.182770","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CertificateReloader","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","SOURCE_LINE":"142"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.182790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CertificateReloader","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","SOURCE_LINE":"142"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185601","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for http:\/\/0.0.0.0:8123","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185651","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for native protocol (tcp): 0.0.0.0:9000","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185697","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185746","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185753","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Ready for connections.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2548"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.221518","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"Poco","MESSAGE":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","SOURCE_LINE":"49"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223495","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCPHandlerFactory","MESSAGE":"TCP Request. Address: 172.16.1.1:57458","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","SOURCE_LINE":"54"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223585","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Client has not sent any data.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","SOURCE_LINE":"330"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223592","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","SOURCE_LINE":"2630"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.449993","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"Poco","MESSAGE":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","SOURCE_LINE":"49"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450067","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCPHandlerFactory","MESSAGE":"TCP Request. Address: 172.16.1.1:57466","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","SOURCE_LINE":"54"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450147","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()","SOURCE_LINE":"1695"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450170","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"Authenticating user 'default' from 172.16.1.1:57466","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","SOURCE_LINE":"364"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450219","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","SOURCE_LINE":"374"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450234","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()","SOURCE_LINE":"534"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450356","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"432"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450388","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"433"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450415","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"434"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.463932","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const","SOURCE_LINE":"664"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464321","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"executeQuery","MESSAGE":"(from 172.16.1.1:57466) (query 1, line 1) SELECT 1 (stage: Complete)","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)","SOURCE_LINE":"237"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464367","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"CancellationChecker","MESSAGE":"Did not add the task because the timeout is 0. Query: SELECT 1","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)","SOURCE_LINE":"77"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464563","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"Planner","MESSAGE":"Query to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()","SOURCE_LINE":"1296"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464667","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"Planner","MESSAGE":"Query from stage FetchColumns to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()","SOURCE_LINE":"1574"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464817","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"QueryMetricLog","MESSAGE":"Scheduling next collecting task for query_id fce58dd8-584b-422e-b892-32498fffbd41 in 999 ms","SOURCE_FILE":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)","SOURCE_LINE":"219"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.465688","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"executeQuery","MESSAGE":"Read 1 rows, 1.00 B in 0.001369 sec., 730.4601899196493 rows\/sec., 730.46 B\/sec.","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)","SOURCE_LINE":"592"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.465806","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"TCPHandler","MESSAGE":"Processed in 0.002027772 sec.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)","SOURCE_LINE":"865"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481884","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Closing connection (open: true, cancelled: false, eof: true)","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","SOURCE_LINE":"456"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481920","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","SOURCE_LINE":"2630"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481958","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; DB::Session::~Session()","SOURCE_LINE":"308"} grep result {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.815901","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"SentryWriter","MESSAGE":"Sending crash reports is disabled","SOURCE_FILE":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)","SOURCE_LINE":"144"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","SOURCE_FILE":"","SOURCE_LINE":"0"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873632","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"starting up","SOURCE_FILE":"","SOURCE_LINE":"0"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873641","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","SOURCE_FILE":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)","SOURCE_LINE":"585"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873735","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Jemalloc","MESSAGE":"Value for background_thread set to true (from true)","SOURCE_FILE":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]","SOURCE_LINE":"32"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877381","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1027"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877418","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1042"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.877455","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Pipe","MESSAGE":"Pipe capacity is 1.00 MiB","SOURCE_FILE":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)","SOURCE_LINE":"131"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878183","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsReader","MESSAGE":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)","SOURCE_LINE":"214"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878344","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/class\/thermal","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()","SOURCE_LINE":"126"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878367","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/block","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()","SOURCE_LINE":"163"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878479","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/devices\/system\/edac","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()","SOURCE_LINE":"190"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.878503","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"AsynchronousMetrics","MESSAGE":"Scanning \/sys\/class\/hwmon","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()","SOURCE_LINE":"220"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.028504","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1349"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.028629","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1403"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032548","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1407"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032616","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"MemoryWorker","MESSAGE":"Starting background memory thread with period of 50ms, using Cgroups as source","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()","SOURCE_LINE":"252"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.032706","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgSchPool","MESSAGE":"Create BackgroundSchedulePool with 512 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.086483","THREAD_NAME":"","THREAD_ID":"14","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"MemoryTracker","MESSAGE":"Correcting the value of global memory tracker from 3.53 MiB to 116.19 MiB","SOURCE_FILE":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)","SOURCE_LINE":"542"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087254","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"rlimit on number of file descriptors is 1048576","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1474"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087291","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"rlimit on number of threads is 18446744073709551615","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1496"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087316","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initializing DateLUT.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1524"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initialized DateLUT with time zone 'Etc\/UTC'.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1526"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087375","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)","SOURCE_LINE":"1324"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087281","THREAD_NAME":"","THREAD_ID":"17","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"CancellationChecker","MESSAGE":"Started worker function","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()","SOURCE_LINE":"97"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087814","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_all_keys' as replica host.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1605"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087848","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Initializing interserver credentials.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"1619"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.087982","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"NamedCollectionsMetadataStorage","MESSAGE":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)","SOURCE_LINE":"627"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088013","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)","SOURCE_LINE":"256"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088026","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from sql","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)","SOURCE_LINE":"276"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.088043","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"FileCacheFactory","MESSAGE":"Will load 0 caches from default cache config","SOURCE_FILE":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)","SOURCE_LINE":"226"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.090409","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loading config '\/etc\/clickhouse-server\/config.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"128"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.090431","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"680"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091652","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091787","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091895","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092137","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_all_keys_json.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.099047","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","SOURCE_LINE":"926"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.099173","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"175"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.107990","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1792"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1819"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Merges and mutations memory limit is set to 15.30 GiB","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1832"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108202","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting max_remote_read_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1874"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Setting max_remote_write_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1875"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108230","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","SOURCE_LINE":"1886"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.108250","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgBufSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110088","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgMBSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.111654","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"BackgroundSchedulePool\/BgDistSchPool","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","SOURCE_LINE":"164"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.114578","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"NamedCollectionFactory","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)","SOURCE_LINE":"267"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.115649","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"193"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.115700","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","SOURCE_LINE":"45"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.116188","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2187"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118842","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loading config '\/etc\/clickhouse-server\/users.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"128"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118861","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"680"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119082","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119124","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","SOURCE_LINE":"709"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119674","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigProcessor","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","SOURCE_LINE":"926"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.119708","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"175"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120173","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","SOURCE_LINE":"193"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120462","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ConfigReloader","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","SOURCE_LINE":"45"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120872","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Access(user directories)","MESSAGE":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)","SOURCE_LINE":"359"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123056","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Access(local_directory)","MESSAGE":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()","SOURCE_LINE":"246"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123081","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Access(local_directory)","MESSAGE":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)","SOURCE_LINE":"186"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123239","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Access(user directories)","MESSAGE":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)","SOURCE_LINE":"397"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsMemoryUsageObserver","MESSAGE":"Started cgroup current memory usage observer thread","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()","SOURCE_LINE":"43"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123724","THREAD_NAME":"","THREAD_ID":"582","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CgroupsMemoryUsageObserver","MESSAGE":"Memory amount initially available to the process is 30.60 GiB","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()","SOURCE_LINE":"67"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.125466","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5940"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.126540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for move operations with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5950"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.127948","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for fetches with num_threads=16, num_tasks=16","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5960"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.128664","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","SOURCE_LINE":"5970"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","SOURCE_LINE":"932"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130702","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","SOURCE_LINE":"932"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130937","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DNSCacheUpdater","MESSAGE":"Update period 15 seconds","SOURCE_FILE":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()","SOURCE_LINE":"49"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130959","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Loading metadata from \/var\/lib\/clickhouse\/","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2315"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131002","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Database disk name: default","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","SOURCE_LINE":"1182"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131009","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Context","MESSAGE":"Database disk name: default, path: \/var\/lib\/clickhouse\/","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","SOURCE_LINE":"1190"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131043","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/system.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131002","THREAD_NAME":"","THREAD_ID":"15","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DNSResolver","MESSAGE":"Updating DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","SOURCE_LINE":"446"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131092","THREAD_NAME":"","THREAD_ID":"15","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DNSResolver","MESSAGE":"Updated DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","SOURCE_LINE":"477"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.132288","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/information_schema.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.133960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"loadSystemDatabase","MESSAGE":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","SOURCE_LINE":"305"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135316","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","SOURCE_LINE":"311"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TablesLoader","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 7.2185e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","SOURCE_LINE":"63"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ReferentialDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135353","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"LoadingDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135410","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","SOURCE_LINE":"693"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135419","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","SOURCE_LINE":"693"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135461","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Atomic database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135470","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: none -> 0","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135479","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #1 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135618","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Wait load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)","SOURCE_LINE":"802"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135672","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Ordinary database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135725","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Ordinary database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135736","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #2 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135803","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135866","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Atomic database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135880","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135915","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.135961","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: 0 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.139132","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_log from query_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.140221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_thread_log from query_thread_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.140682","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.part_log from part_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141008","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.trace_log from trace_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141281","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.crash_log from crash_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141448","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.text_log from text_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.141703","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.metric_log from metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146046","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.latency_log from latency_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146341","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.error_log from error_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146584","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146610","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146627","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.s3queue_log from s3queue_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146814","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.146830","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.asynchronous_metric_log from asynchronous_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147010","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.opentelemetry_span_log from opentelemetry_span_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_views_log from query_views_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147574","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147603","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.session_log since corresponding section 'session_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147615","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]","SOURCE_LINE":"153"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147626","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.processors_profile_log from processors_profile_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.147926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.asynchronous_insert_log from asynchronous_insert_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.backup_log from backup_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148572","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.blob_storage_log from blob_storage_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.148807","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"SystemLog","MESSAGE":"Creating system.query_metric_log from query_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]","SOURCE_LINE":"158"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.153508","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.153540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154821","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154834","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154855","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154860","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154881","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154887","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154906","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154911","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154949","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154962","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155060","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155067","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155112","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155118","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155170","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155216","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155222","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155254","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155260","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155343","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155379","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155384","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155421","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155428","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155547","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155588","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155594","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155619","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155629","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155660","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155683","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155688","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155708","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155733","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155738","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155760","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155768","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155795","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155830","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155854","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155860","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155922","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155928","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155955","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156141","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156152","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156181","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156187","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156209","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156213","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156237","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156242","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156271","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156276","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156348","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156357","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156395","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156400","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156484","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156493","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156544","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156550","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156630","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156636","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156712","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156723","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156762","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156767","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156891","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156898","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156928","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156934","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156954","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156958","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156984","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156994","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157028","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157056","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157061","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157080","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157085","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157135","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157186","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157219","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157417","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157431","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157534","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157542","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157582","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157587","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157607","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157614","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157659","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157665","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157707","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157737","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157742","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157761","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157766","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157914","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158062","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158075","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158141","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158263","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158270","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158412","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158430","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158508","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158516","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158555","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158561","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158706","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158716","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158776","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158782","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158827","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158833","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158868","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158873","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158912","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158918","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159051","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159061","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159117","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159123","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159172","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159222","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159283","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159289","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159336","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159399","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159406","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159501","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159510","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159546","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159551","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159582","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159587","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159624","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159633","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159669","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159674","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159716","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159780","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159786","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159831","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159858","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159863","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159905","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159910","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159942","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159947","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160002","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160008","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160081","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160088","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160116","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160121","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160157","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160190","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160196","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160230","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160235","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160261","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160266","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160288","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160293","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160354","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160365","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160420","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160437","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160467","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160472","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160549","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160556","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.167597","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseCatalog","MESSAGE":"Found 0 partially dropped tables. Will load them and retry removal.","SOURCE_FILE":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()","SOURCE_LINE":"1050"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168775","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (default)","MESSAGE":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","SOURCE_LINE":"311"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168788","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TablesLoader","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 2.616e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","SOURCE_LINE":"63"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168800","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ReferentialDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"LoadingDeps","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","SOURCE_LINE":"722"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168832","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"loadMetadata","MESSAGE":"Start asynchronous loading of databases","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)","SOURCE_LINE":"268"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168841","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168846","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Schedule load job 'startup Atomic database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","SOURCE_LINE":"347"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168851","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: none -> 2","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168856","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #1 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168975","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"UserDefinedSQLObjectsLoaderFromDisk","MESSAGE":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","SOURCE_LINE":"131"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168990","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"UserDefinedSQLObjectsLoaderFromDisk","MESSAGE":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","SOURCE_LINE":"135"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"WorkloadEntityDiskStorage","MESSAGE":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","SOURCE_LINE":"129"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169031","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"WorkloadEntityDiskStorage","MESSAGE":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","SOURCE_LINE":"133"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169046","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Loaded metadata.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2408"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169040","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Ordinary database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169095","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Ordinary database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169105","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Spawn loader worker #2 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","SOURCE_LINE":"883"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169137","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"bool DB::(anonymous namespace)::checkPermissionsImpl()","MESSAGE":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()","SOURCE_LINE":"216"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169157","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Tasks stats provider: procfs","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2428"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169179","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Execute load job 'startup Atomic database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"934"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169190","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Finish load job 'startup Atomic database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","SOURCE_LINE":"606"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169197","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169204","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"440"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169220","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"DatabaseAtomic (system)","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","SOURCE_LINE":"451"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169283","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","SOURCE_LINE":"916"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169359","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"AsyncLoader","MESSAGE":"Change current priority: 2 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","SOURCE_LINE":"850"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169474","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","SOURCE_LINE":"37"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169517","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","SOURCE_LINE":"48"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169524","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"MySQLHandlerFactory","MESSAGE":"Generating new RSA key pair.","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()","SOURCE_LINE":"107"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.182770","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CertificateReloader","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","SOURCE_LINE":"142"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.182790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"CertificateReloader","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","SOURCE_LINE":"142"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185601","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for http:\/\/0.0.0.0:8123","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185651","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for native protocol (tcp): 0.0.0.0:9000","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185697","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185746","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2544"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.185753","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"Application","MESSAGE":"Ready for connections.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","SOURCE_LINE":"2548"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.221518","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"Poco","MESSAGE":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","SOURCE_LINE":"49"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223495","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCPHandlerFactory","MESSAGE":"TCP Request. Address: 172.16.1.1:57458","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","SOURCE_LINE":"54"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223585","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Information","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Client has not sent any data.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","SOURCE_LINE":"330"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.223592","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","SOURCE_LINE":"2630"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.449993","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"Poco","MESSAGE":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","SOURCE_LINE":"49"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450067","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCPHandlerFactory","MESSAGE":"TCP Request. Address: 172.16.1.1:57466","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","SOURCE_LINE":"54"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450147","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()","SOURCE_LINE":"1695"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450170","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"Authenticating user 'default' from 172.16.1.1:57466","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","SOURCE_LINE":"364"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450219","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","SOURCE_LINE":"374"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450234","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()","SOURCE_LINE":"534"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450356","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"432"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450388","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"433"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.450415","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"ContextAccess (default)","MESSAGE":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","SOURCE_LINE":"434"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.463932","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const","SOURCE_LINE":"664"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464321","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"executeQuery","MESSAGE":"(from 172.16.1.1:57466) (query 1, line 1) SELECT 1 (stage: Complete)","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)","SOURCE_LINE":"237"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464367","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"CancellationChecker","MESSAGE":"Did not add the task because the timeout is 0. Query: SELECT 1","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)","SOURCE_LINE":"77"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464563","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"Planner","MESSAGE":"Query to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()","SOURCE_LINE":"1296"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464667","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"Planner","MESSAGE":"Query from stage FetchColumns to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()","SOURCE_LINE":"1574"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.464817","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"QueryMetricLog","MESSAGE":"Scheduling next collecting task for query_id fce58dd8-584b-422e-b892-32498fffbd41 in 999 ms","SOURCE_FILE":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)","SOURCE_LINE":"219"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.465688","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"executeQuery","MESSAGE":"Read 1 rows, 1.00 B in 0.001369 sec., 730.4601899196493 rows\/sec., 730.46 B\/sec.","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)","SOURCE_LINE":"592"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.465806","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"fce58dd8-584b-422e-b892-32498fffbd41","LOGGER_NAME":"TCPHandler","MESSAGE":"Processed in 0.002027772 sec.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)","SOURCE_LINE":"865"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481884","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Closing connection (open: true, cancelled: false, eof: true)","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","SOURCE_LINE":"456"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481920","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCPHandler","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","SOURCE_LINE":"2630"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.481958","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","QUERY_ID":"","LOGGER_NAME":"TCP-Session-43681550-0b31-4026-b226-a4b02b99ad6c","MESSAGE":"43681550-0b31-4026-b226-a4b02b99ad6c Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; DB::Session::~Session()","SOURCE_LINE":"308"} Executing query SELECT 1 on node_some_keys Executing query SELECT 1 on node_some_keys Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (86, '0x1ce4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (86, '0x1ce4') on node grep in log called grep in log called run container_id:rootteststructuredloggingjson-gw8-node_some_keys-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true'] run container_id:rootteststructuredloggingjson-gw8-node_some_keys-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true'] Command:[docker exec rootteststructuredloggingjson-gw8-node_some_keys-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true] Command:[docker exec rootteststructuredloggingjson-gw8-node_some_keys-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true] Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.792807","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Sending crash reports is disabled","SOURCE_FILE":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.792807","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Sending crash reports is disabled","SOURCE_FILE":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.859991","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","SOURCE_FILE":""} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.859991","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","SOURCE_FILE":""} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"starting up","SOURCE_FILE":""} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"starting up","SOURCE_FILE":""} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860181","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","SOURCE_FILE":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860181","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","SOURCE_FILE":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860285","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Value for background_thread set to true (from true)","SOURCE_FILE":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860285","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Value for background_thread set to true (from true)","SOURCE_FILE":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865758","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865758","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865798","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865798","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865836","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Pipe capacity is 1.00 MiB","SOURCE_FILE":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865836","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Pipe capacity is 1.00 MiB","SOURCE_FILE":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.872779","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.872779","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873382","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/class\/thermal","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873382","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/class\/thermal","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/block","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/block","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/devices\/system\/edac","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/devices\/system\/edac","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873564","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/class\/hwmon","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873564","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/class\/hwmon","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.035554","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.035554","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.035692","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.035692","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039521","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039521","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039553","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Starting background memory thread with period of 50ms, using Cgroups as source","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039553","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Starting background memory thread with period of 50ms, using Cgroups as source","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039620","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 512 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039620","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 512 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091817","THREAD_NAME":"","THREAD_ID":"16","LEVEL":"Information","MESSAGE":"Correcting the value of global memory tracker from 3.52 MiB to 116.02 MiB","SOURCE_FILE":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091817","THREAD_NAME":"","THREAD_ID":"16","LEVEL":"Information","MESSAGE":"Correcting the value of global memory tracker from 3.52 MiB to 116.02 MiB","SOURCE_FILE":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092844","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"rlimit on number of file descriptors is 1048576","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092844","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"rlimit on number of file descriptors is 1048576","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092880","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"rlimit on number of threads is 18446744073709551615","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092880","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"rlimit on number of threads is 18446744073709551615","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092894","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Initializing DateLUT.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092894","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Initializing DateLUT.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092901","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Initialized DateLUT with time zone 'Etc\/UTC'.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092901","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Initialized DateLUT with time zone 'Etc\/UTC'.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092891","THREAD_NAME":"","THREAD_ID":"14","LEVEL":"Trace","MESSAGE":"Started worker function","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092891","THREAD_NAME":"","THREAD_ID":"14","LEVEL":"Trace","MESSAGE":"Started worker function","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093477","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_some_keys' as replica host.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093477","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_some_keys' as replica host.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093503","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Initializing interserver credentials.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093503","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Initializing interserver credentials.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093624","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093624","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093666","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from sql","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093666","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from sql","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093681","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Will load 0 caches from default cache config","SOURCE_FILE":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093681","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Will load 0 caches from default cache config","SOURCE_FILE":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.096011","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loading config '\/etc\/clickhouse-server\/config.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.096011","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loading config '\/etc\/clickhouse-server\/config.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.096035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.096035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097220","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097220","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097354","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097354","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097452","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097452","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097678","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_some_keys_json.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097678","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_some_keys_json.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.103828","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.103828","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.103926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.103926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110155","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110155","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110171","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110171","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Merges and mutations memory limit is set to 15.30 GiB","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Merges and mutations memory limit is set to 15.30 GiB","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110293","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting max_remote_read_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110293","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting max_remote_read_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110314","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting max_remote_write_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110314","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting max_remote_write_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110334","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110334","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.112380","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.112380","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.114163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.114163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118413","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118413","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118476","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118476","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118863","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118863","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loading config '\/etc\/clickhouse-server\/users.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loading config '\/etc\/clickhouse-server\/users.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120734","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120734","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120883","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120883","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120922","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120922","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.121933","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.121933","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122328","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122328","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122345","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122345","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122821","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122821","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122841","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122841","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123107","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123107","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123260","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123260","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123532","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Started cgroup current memory usage observer thread","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123532","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Started cgroup current memory usage observer thread","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123921","THREAD_NAME":"","THREAD_ID":"582","LEVEL":"Information","MESSAGE":"Memory amount initially available to the process is 30.60 GiB","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123921","THREAD_NAME":"","THREAD_ID":"582","LEVEL":"Information","MESSAGE":"Memory amount initially available to the process is 30.60 GiB","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.125607","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.125607","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.126525","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for move operations with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.126525","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for move operations with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.127980","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for fetches with num_threads=16, num_tasks=16","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.127980","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for fetches with num_threads=16, num_tasks=16","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.128728","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.128728","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131093","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Update period 15 seconds","SOURCE_FILE":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131093","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Update period 15 seconds","SOURCE_FILE":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131120","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading metadata from \/var\/lib\/clickhouse\/","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131120","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading metadata from \/var\/lib\/clickhouse\/","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131166","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Database disk name: default","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131166","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Database disk name: default","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Database disk name: default, path: \/var\/lib\/clickhouse\/","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Database disk name: default, path: \/var\/lib\/clickhouse\/","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/system.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/system.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131354","THREAD_NAME":"","THREAD_ID":"18","LEVEL":"Debug","MESSAGE":"Updating DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131354","THREAD_NAME":"","THREAD_ID":"18","LEVEL":"Debug","MESSAGE":"Updating DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131393","THREAD_NAME":"","THREAD_ID":"18","LEVEL":"Debug","MESSAGE":"Updated DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131393","THREAD_NAME":"","THREAD_ID":"18","LEVEL":"Debug","MESSAGE":"Updated DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.132284","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/information_schema.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.132284","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/information_schema.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.133295","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.133295","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134347","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 5.9391e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134347","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 5.9391e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134357","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134357","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134368","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134368","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134407","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134407","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134419","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134419","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134432","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134432","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134441","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Atomic database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134441","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Atomic database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Change current priority: none -> 0","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Change current priority: none -> 0","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134457","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Spawn loader worker #1 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134457","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Spawn loader worker #1 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134533","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Wait load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134533","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Wait load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134619","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Ordinary database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134619","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Ordinary database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134704","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Ordinary database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134704","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Ordinary database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134714","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Spawn loader worker #2 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134714","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Spawn loader worker #2 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134789","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134789","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134857","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134857","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134954","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Atomic database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134954","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Atomic database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134972","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134972","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134984","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Change current priority: 0 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134984","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Change current priority: 0 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134991","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_log from query_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134991","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_log from query_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.136282","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_thread_log from query_thread_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.136282","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_thread_log from query_thread_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.136865","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.part_log from part_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.136865","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.part_log from part_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137214","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.trace_log from trace_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137214","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.trace_log from trace_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137498","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.crash_log from crash_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137498","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.crash_log from crash_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137649","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.text_log from text_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137649","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.text_log from text_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137914","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.metric_log from metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137914","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.metric_log from metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.142622","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.latency_log from latency_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.142622","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.latency_log from latency_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.142919","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.error_log from error_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.142919","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.error_log from error_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143131","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143131","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143149","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143149","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.s3queue_log from s3queue_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.s3queue_log from s3queue_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143371","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143371","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143390","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.asynchronous_metric_log from asynchronous_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143390","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.asynchronous_metric_log from asynchronous_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143679","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.opentelemetry_span_log from opentelemetry_span_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143679","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.opentelemetry_span_log from opentelemetry_span_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_views_log from query_views_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_views_log from query_views_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144199","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144199","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.session_log since corresponding section 'session_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.session_log since corresponding section 'session_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144233","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144233","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144243","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.processors_profile_log from processors_profile_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144243","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.processors_profile_log from processors_profile_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144600","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.asynchronous_insert_log from asynchronous_insert_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144600","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.asynchronous_insert_log from asynchronous_insert_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144857","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.backup_log from backup_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144857","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.backup_log from backup_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.145119","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.blob_storage_log from blob_storage_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.145119","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.blob_storage_log from blob_storage_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.145364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_metric_log from query_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.145364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_metric_log from query_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.151781","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.151781","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.151805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.151805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154272","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154272","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154281","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154281","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154333","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154333","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154384","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154384","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154392","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154392","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154460","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154460","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154469","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154469","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154615","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154615","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154628","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154628","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154702","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154702","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154792","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154792","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154801","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154801","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154869","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154869","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154880","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154880","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154939","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154939","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154949","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154949","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155011","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155011","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155020","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155020","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155070","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155070","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155080","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155080","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155158","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155158","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155216","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155216","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155225","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155225","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155282","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155282","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155291","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155291","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155344","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155344","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155353","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155353","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155391","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155391","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155400","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155400","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155485","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155485","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155499","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155499","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155538","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155538","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155546","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155546","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155589","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155589","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155598","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155598","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155634","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155634","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155642","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155642","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155673","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155673","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155681","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155681","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155729","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155729","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155738","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155738","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155774","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155774","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155782","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155782","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156030","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156030","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156045","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156045","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156089","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156089","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156103","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156103","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156245","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156245","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156255","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156255","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156292","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156292","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156312","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156312","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156338","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156338","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156343","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156343","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156369","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156369","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156396","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156396","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156401","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156401","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156460","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156460","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156472","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156472","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156528","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156528","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156536","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156536","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156599","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156599","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156608","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156608","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156661","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156661","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156732","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156732","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156739","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156739","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156811","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156811","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156820","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156820","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156890","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156890","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156899","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156899","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156989","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156989","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157000","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157000","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157028","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157028","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157051","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157051","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157055","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157055","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157078","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157078","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157086","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157086","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157117","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157117","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157123","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157123","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157143","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157143","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157169","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157169","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157228","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157228","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157261","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157261","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157269","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157269","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157295","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157295","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157315","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157315","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157454","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157454","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157461","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157461","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157542","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157542","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157548","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157548","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157580","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157580","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157585","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157585","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157604","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157604","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157616","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157616","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157658","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157658","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157664","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157664","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157705","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157705","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157732","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157732","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157736","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157736","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157753","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157753","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157757","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157757","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157930","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157930","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157946","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157946","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158121","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158121","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158133","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158133","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158200","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158200","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158207","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158207","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158430","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158430","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158436","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158436","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158483","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158483","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158491","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158491","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158530","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158530","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158536","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158536","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158675","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158675","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158683","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158683","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158736","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158736","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158746","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158746","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158789","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158789","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158794","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158794","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158833","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158833","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158871","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158871","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158876","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158876","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159005","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159005","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159013","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159013","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159068","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159068","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159073","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159073","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159125","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159125","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159131","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159131","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159172","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159172","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159177","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159177","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159231","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159231","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159237","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159237","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159265","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159265","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159270","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159270","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159334","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159334","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159340","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159340","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159378","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159378","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159383","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159383","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159407","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159407","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159494","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159494","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159559","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159559","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159566","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159566","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159601","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159601","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159606","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159606","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159661","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159661","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159673","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159673","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159795","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159795","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159845","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159845","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159851","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159851","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159881","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159881","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159886","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159886","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159930","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159930","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159935","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159935","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159966","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159966","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159971","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159971","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160027","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160027","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160107","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160107","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160115","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160115","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160174","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160174","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160206","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160206","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160215","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160215","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160247","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160247","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160252","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160252","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160275","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160275","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160280","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160280","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160315","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160315","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160320","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160320","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160359","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160359","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160391","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160391","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160396","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160396","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160423","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160423","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160428","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160428","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160507","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160507","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168972","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Found 0 partially dropped tables. Will load them and retry removal.","SOURCE_FILE":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168972","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Found 0 partially dropped tables. Will load them and retry removal.","SOURCE_FILE":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169993","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169993","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170010","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 3.4274e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170010","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 3.4274e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170018","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170018","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170057","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Start asynchronous loading of databases","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170057","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Start asynchronous loading of databases","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170069","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170069","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170076","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Atomic database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170076","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Atomic database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170084","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Change current priority: none -> 2","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170084","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Change current priority: none -> 2","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170091","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Spawn loader worker #1 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170091","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Spawn loader worker #1 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171740","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Ordinary database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171740","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Ordinary database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171798","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Ordinary database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171798","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Ordinary database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171812","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Spawn loader worker #2 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171812","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Spawn loader worker #2 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171815","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171815","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171840","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171840","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171891","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171891","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171899","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Atomic database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171899","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Atomic database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171906","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171906","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171911","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Atomic database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171911","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Atomic database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171923","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded metadata.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171923","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded metadata.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171932","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171932","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172069","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172069","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172033","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172033","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172106","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Tasks stats provider: procfs","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172106","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Tasks stats provider: procfs","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172104","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","MESSAGE":"Change current priority: 2 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172104","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","MESSAGE":"Change current priority: 2 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172190","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172190","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172205","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172205","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172492","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172492","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172541","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172541","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172549","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Generating new RSA key pair.","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172549","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Generating new RSA key pair.","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.226972","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.226972","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.226997","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.226997","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for http:\/\/0.0.0.0:8123","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for http:\/\/0.0.0.0:8123","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230720","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for native protocol (tcp): 0.0.0.0:9000","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230720","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for native protocol (tcp): 0.0.0.0:9000","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230771","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230771","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230823","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230823","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230845","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Ready for connections.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230845","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Ready for connections.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231790","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","MESSAGE":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231790","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","MESSAGE":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231896","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"TCP Request. Address: 172.16.1.1:56132","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231896","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"TCP Request. Address: 172.16.1.1:56132","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231977","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Information","MESSAGE":"Client has not sent any data.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231977","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Information","MESSAGE":"Client has not sent any data.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231988","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231988","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628542","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","MESSAGE":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628542","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","MESSAGE":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628643","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"TCP Request. Address: 172.16.1.1:56136","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628643","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"TCP Request. Address: 172.16.1.1:56136","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628738","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628738","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628764","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Authenticating user 'default' from 172.16.1.1:56136","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628764","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Authenticating user 'default' from 172.16.1.1:56136","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628811","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628811","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628837","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628837","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628971","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628971","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.629035","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.629035","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.629069","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.629069","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638472","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638472","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638829","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"(from 172.16.1.1:56136) (query 1, line 1) SELECT 1 (stage: Complete)","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638829","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"(from 172.16.1.1:56136) (query 1, line 1) SELECT 1 (stage: Complete)","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638875","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Did not add the task because the timeout is 0. Query: SELECT 1","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638875","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Did not add the task because the timeout is 0. Query: SELECT 1","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639064","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Query to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639064","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Query to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639145","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Query from stage FetchColumns to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639145","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Query from stage FetchColumns to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639280","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Scheduling next collecting task for query_id 0047e1e3-8089-4426-996a-de12bf1f6cf9 in 999 ms","SOURCE_FILE":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639280","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Scheduling next collecting task for query_id 0047e1e3-8089-4426-996a-de12bf1f6cf9 in 999 ms","SOURCE_FILE":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.640101","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Read 1 rows, 1.00 B in 0.001285 sec., 778.2101167315176 rows\/sec., 778.21 B\/sec.","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.640101","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Read 1 rows, 1.00 B in 0.001285 sec., 778.2101167315176 rows\/sec., 778.21 B\/sec.","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.640237","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Processed in 0.001937624 sec.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.640237","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Processed in 0.001937624 sec.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650021","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Closing connection (open: true, cancelled: false, eof: true)","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650021","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Closing connection (open: true, cancelled: false, eof: true)","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650047","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650047","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650070","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; DB::Session::~Session()"} Stdout:{"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650070","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; DB::Session::~Session()"} grep result {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.792807","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Sending crash reports is disabled","SOURCE_FILE":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.859991","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","SOURCE_FILE":""} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"starting up","SOURCE_FILE":""} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860181","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","SOURCE_FILE":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860285","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Value for background_thread set to true (from true)","SOURCE_FILE":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865758","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865798","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865836","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Pipe capacity is 1.00 MiB","SOURCE_FILE":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.872779","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873382","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/class\/thermal","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/block","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/devices\/system\/edac","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873564","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/class\/hwmon","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.035554","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.035692","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039521","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039553","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Starting background memory thread with period of 50ms, using Cgroups as source","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039620","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 512 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091817","THREAD_NAME":"","THREAD_ID":"16","LEVEL":"Information","MESSAGE":"Correcting the value of global memory tracker from 3.52 MiB to 116.02 MiB","SOURCE_FILE":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092844","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"rlimit on number of file descriptors is 1048576","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092880","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"rlimit on number of threads is 18446744073709551615","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092894","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Initializing DateLUT.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092901","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Initialized DateLUT with time zone 'Etc\/UTC'.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092891","THREAD_NAME":"","THREAD_ID":"14","LEVEL":"Trace","MESSAGE":"Started worker function","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093477","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_some_keys' as replica host.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093503","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Initializing interserver credentials.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093624","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093666","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from sql","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093681","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Will load 0 caches from default cache config","SOURCE_FILE":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.096011","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loading config '\/etc\/clickhouse-server\/config.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.096035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097220","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097354","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097452","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097678","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_some_keys_json.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.103828","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.103926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110155","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110171","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Merges and mutations memory limit is set to 15.30 GiB","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110293","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting max_remote_read_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110314","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting max_remote_write_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110334","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.112380","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.114163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118413","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118476","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118863","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loading config '\/etc\/clickhouse-server\/users.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120734","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120883","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120922","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.121933","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122328","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122345","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122821","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122841","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123107","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123260","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123532","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Started cgroup current memory usage observer thread","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123921","THREAD_NAME":"","THREAD_ID":"582","LEVEL":"Information","MESSAGE":"Memory amount initially available to the process is 30.60 GiB","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.125607","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.126525","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for move operations with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.127980","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for fetches with num_threads=16, num_tasks=16","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.128728","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131093","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Update period 15 seconds","SOURCE_FILE":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131120","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading metadata from \/var\/lib\/clickhouse\/","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131166","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Database disk name: default","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Database disk name: default, path: \/var\/lib\/clickhouse\/","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/system.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131354","THREAD_NAME":"","THREAD_ID":"18","LEVEL":"Debug","MESSAGE":"Updating DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131393","THREAD_NAME":"","THREAD_ID":"18","LEVEL":"Debug","MESSAGE":"Updated DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.132284","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/information_schema.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.133295","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134347","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 5.9391e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134357","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134368","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134407","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134419","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134432","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134441","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Atomic database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Change current priority: none -> 0","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134457","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Spawn loader worker #1 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134533","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Wait load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134619","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Ordinary database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134704","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Ordinary database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134714","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Spawn loader worker #2 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134789","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134857","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134954","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Atomic database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134972","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134984","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Change current priority: 0 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134991","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_log from query_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.136282","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_thread_log from query_thread_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.136865","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.part_log from part_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137214","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.trace_log from trace_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137498","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.crash_log from crash_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137649","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.text_log from text_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137914","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.metric_log from metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.142622","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.latency_log from latency_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.142919","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.error_log from error_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143131","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143149","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.s3queue_log from s3queue_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143371","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143390","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.asynchronous_metric_log from asynchronous_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143679","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.opentelemetry_span_log from opentelemetry_span_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_views_log from query_views_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144199","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.session_log since corresponding section 'session_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144233","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144243","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.processors_profile_log from processors_profile_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144600","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.asynchronous_insert_log from asynchronous_insert_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144857","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.backup_log from backup_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.145119","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.blob_storage_log from blob_storage_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.145364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_metric_log from query_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.151781","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.151805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154272","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154281","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154333","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154384","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154392","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154460","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154469","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154615","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154628","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154702","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154792","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154801","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154869","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154880","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154939","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154949","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155011","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155020","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155070","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155080","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155158","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155216","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155225","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155282","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155291","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155344","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155353","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155391","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155400","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155485","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155499","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155538","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155546","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155589","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155598","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155634","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155642","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155673","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155681","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155729","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155738","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155774","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155782","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156030","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156045","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156089","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156103","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156245","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156255","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156292","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156312","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156338","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156343","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156369","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156396","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156401","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156460","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156472","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156528","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156536","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156599","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156608","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156661","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156732","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156739","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156811","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156820","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156890","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156899","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156989","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157000","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157028","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157051","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157055","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157078","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157086","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157117","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157123","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157143","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157169","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157228","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157261","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157269","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157295","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157315","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157454","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157461","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157542","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157548","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157580","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157585","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157604","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157616","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157658","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157664","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157705","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157732","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157736","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157753","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157757","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157930","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157946","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158121","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158133","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158200","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158207","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158430","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158436","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158483","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158491","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158530","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158536","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158675","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158683","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158736","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158746","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158789","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158794","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158833","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158871","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158876","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159005","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159013","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159068","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159073","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159125","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159131","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159172","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159177","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159231","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159237","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159265","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159270","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159334","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159340","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159378","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159383","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159407","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159494","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159559","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159566","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159601","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159606","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159661","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159673","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159795","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159845","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159851","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159881","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159886","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159930","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159935","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159966","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159971","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160027","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160107","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160115","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160174","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160206","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160215","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160247","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160252","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160275","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160280","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160315","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160320","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160359","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160391","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160396","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160423","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160428","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160507","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168972","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Found 0 partially dropped tables. Will load them and retry removal.","SOURCE_FILE":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169993","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170010","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 3.4274e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170018","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170057","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Start asynchronous loading of databases","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170069","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170076","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Atomic database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170084","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Change current priority: none -> 2","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170091","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Spawn loader worker #1 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171740","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Ordinary database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171798","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Ordinary database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171812","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Spawn loader worker #2 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171815","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171840","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171891","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171899","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Atomic database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171906","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171911","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Atomic database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171923","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded metadata.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171932","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172069","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172033","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172106","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Tasks stats provider: procfs","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172104","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","MESSAGE":"Change current priority: 2 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172190","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172205","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172492","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172541","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172549","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Generating new RSA key pair.","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.226972","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.226997","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for http:\/\/0.0.0.0:8123","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230720","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for native protocol (tcp): 0.0.0.0:9000","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230771","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230823","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230845","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Ready for connections.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231790","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","MESSAGE":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231896","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"TCP Request. Address: 172.16.1.1:56132","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231977","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Information","MESSAGE":"Client has not sent any data.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231988","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628542","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","MESSAGE":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628643","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"TCP Request. Address: 172.16.1.1:56136","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628738","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628764","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Authenticating user 'default' from 172.16.1.1:56136","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628811","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628837","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628971","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.629035","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.629069","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638472","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638829","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"(from 172.16.1.1:56136) (query 1, line 1) SELECT 1 (stage: Complete)","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638875","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Did not add the task because the timeout is 0. Query: SELECT 1","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639064","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Query to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639145","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Query from stage FetchColumns to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639280","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Scheduling next collecting task for query_id 0047e1e3-8089-4426-996a-de12bf1f6cf9 in 999 ms","SOURCE_FILE":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.640101","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Read 1 rows, 1.00 B in 0.001285 sec., 778.2101167315176 rows\/sec., 778.21 B\/sec.","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.640237","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Processed in 0.001937624 sec.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650021","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Closing connection (open: true, cancelled: false, eof: true)","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650047","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650070","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; DB::Session::~Session()"} grep result {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.792807","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Sending crash reports is disabled","SOURCE_FILE":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.859991","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","SOURCE_FILE":""} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"starting up","SOURCE_FILE":""} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860181","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","SOURCE_FILE":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.860285","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Value for background_thread set to true (from true)","SOURCE_FILE":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865758","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865798","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.865836","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Pipe capacity is 1.00 MiB","SOURCE_FILE":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.872779","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873382","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/class\/thermal","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/block","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873540","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/devices\/system\/edac","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()"} {"DATE_TIME_UTC":"2025-04-02T04:15:42Z","DATE_TIME":"1743567342.873564","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Scanning \/sys\/class\/hwmon","SOURCE_FILE":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.035554","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.035692","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039521","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039553","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Starting background memory thread with period of 50ms, using Cgroups as source","SOURCE_FILE":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.039620","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 512 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.091817","THREAD_NAME":"","THREAD_ID":"16","LEVEL":"Information","MESSAGE":"Correcting the value of global memory tracker from 3.52 MiB to 116.02 MiB","SOURCE_FILE":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092844","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"rlimit on number of file descriptors is 1048576","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092880","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"rlimit on number of threads is 18446744073709551615","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092894","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Initializing DateLUT.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092901","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Initialized DateLUT with time zone 'Etc\/UTC'.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092891","THREAD_NAME":"","THREAD_ID":"14","LEVEL":"Trace","MESSAGE":"Started worker function","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.092960","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093477","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_some_keys' as replica host.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093503","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Initializing interserver credentials.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093624","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093666","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from sql","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.093681","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Will load 0 caches from default cache config","SOURCE_FILE":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.096011","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loading config '\/etc\/clickhouse-server\/config.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.096035","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097220","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097354","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097452","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.097678","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_some_keys_json.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.103828","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.103926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110155","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110171","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110178","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Merges and mutations memory limit is set to 15.30 GiB","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110293","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting max_remote_read_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110314","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Setting max_remote_write_network_bandwidth_for_server was set to 0","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","SOURCE_FILE":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.110334","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.112380","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.114163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Create BackgroundSchedulePool with 16 threads","SOURCE_FILE":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118413","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Test","MESSAGE":"Loaded 0 collections from config","SOURCE_FILE":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118476","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.118863","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120713","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loading config '\/etc\/clickhouse-server\/users.xml'","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120734","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120883","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.120922","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.121933","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122328","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","SOURCE_FILE":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122345","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122821","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.122841","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Config reload interval set to 2000ms","SOURCE_FILE":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123107","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123260","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","SOURCE_FILE":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123532","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Started cgroup current memory usage observer thread","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.123921","THREAD_NAME":"","THREAD_ID":"582","LEVEL":"Information","MESSAGE":"Memory amount initially available to the process is 30.60 GiB","SOURCE_FILE":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.125607","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.126525","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for move operations with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.127980","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for fetches with num_threads=16, num_tasks=16","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.128728","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.130790","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Warning","MESSAGE":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","SOURCE_FILE":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131093","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Update period 15 seconds","SOURCE_FILE":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131120","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading metadata from \/var\/lib\/clickhouse\/","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131166","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Database disk name: default","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Database disk name: default, path: \/var\/lib\/clickhouse\/","SOURCE_FILE":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131217","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/system.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131354","THREAD_NAME":"","THREAD_ID":"18","LEVEL":"Debug","MESSAGE":"Updating DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.131393","THREAD_NAME":"","THREAD_ID":"18","LEVEL":"Debug","MESSAGE":"Updated DNS cache","SOURCE_FILE":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.132284","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/information_schema.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.133295","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134347","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 5.9391e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134357","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134368","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134407","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134419","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134432","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134441","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Atomic database system' into ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134450","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Change current priority: none -> 0","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134457","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Spawn loader worker #1 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134533","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Wait load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134619","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Ordinary database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134704","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Ordinary database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134714","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Spawn loader worker #2 in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134789","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Atomic database system' in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134857","THREAD_NAME":"","THREAD_ID":"648","LEVEL":"Debug","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134954","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Atomic database system' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134972","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Stop worker in ForegroundLoad","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134984","THREAD_NAME":"","THREAD_ID":"647","LEVEL":"Debug","MESSAGE":"Change current priority: 0 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.134991","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_log from query_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.136282","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_thread_log from query_thread_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.136865","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.part_log from part_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137214","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.trace_log from trace_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137498","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.crash_log from crash_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137649","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.text_log from text_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.137914","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.metric_log from metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.142622","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.latency_log from latency_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.142919","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.error_log from error_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143131","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143149","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143165","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.s3queue_log from s3queue_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143371","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143390","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.asynchronous_metric_log from asynchronous_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143679","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.opentelemetry_span_log from opentelemetry_span_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.143926","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_views_log from query_views_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144199","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.session_log since corresponding section 'session_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144233","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144243","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.processors_profile_log from processors_profile_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144600","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.asynchronous_insert_log from asynchronous_insert_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.144857","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.backup_log from backup_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.145119","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.blob_storage_log from blob_storage_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.145364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Creating system.query_metric_log from query_metric_log","SOURCE_FILE":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.151781","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.151805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154224","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154238","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154272","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154281","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154333","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154346","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154384","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154392","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154460","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154469","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154615","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154628","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154702","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154792","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154801","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154869","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154880","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154939","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.154949","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155011","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155020","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155070","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155080","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155158","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155216","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155225","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155282","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155291","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155344","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155353","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155391","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155400","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155485","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155499","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155538","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155546","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155589","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155598","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155634","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155642","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155673","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155681","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155729","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155738","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155774","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.155782","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156030","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156045","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156089","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156103","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156245","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156255","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156292","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156312","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156338","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156343","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156369","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156396","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156401","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156460","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156472","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156528","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156536","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156599","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156608","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156661","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156732","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156739","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156811","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156820","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156890","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156899","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.156989","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157000","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157028","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157051","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157055","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157078","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157086","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157117","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157123","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157143","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157163","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157169","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157221","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157228","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157261","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157269","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157295","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157315","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157454","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157461","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157542","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157548","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157580","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157585","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157604","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157616","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157658","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157664","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157705","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157711","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157732","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157736","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157753","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157757","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157930","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.157946","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158121","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158133","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158200","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158207","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158322","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158329","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158430","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158436","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158483","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158491","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158530","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158536","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158675","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158683","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158736","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158746","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158789","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158794","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158825","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158833","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158871","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.158876","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159005","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159013","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159068","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159073","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159125","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159131","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159172","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159177","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159231","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159237","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159265","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159270","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159334","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159340","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159378","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159383","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159407","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159411","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159494","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159559","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159566","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159601","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159606","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159661","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159673","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159795","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159805","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159845","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159851","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159881","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159886","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159930","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159935","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159966","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.159971","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160027","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160033","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160107","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160115","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160142","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160147","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160174","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160179","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160206","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160215","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160247","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160252","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160275","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160280","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160315","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160320","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160359","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160364","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160391","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160396","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160423","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160428","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160507","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.160513","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.168972","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Found 0 partially dropped tables. Will load them and retry removal.","SOURCE_FILE":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.169993","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","SOURCE_FILE":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170010","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Parsed metadata of 0 tables in 1 databases in 3.4274e-05 sec","SOURCE_FILE":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170018","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170024","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"No tables","SOURCE_FILE":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170057","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Start asynchronous loading of databases","SOURCE_FILE":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170069","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170076","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Schedule load job 'startup Atomic database default' into BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170084","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Change current priority: none -> 2","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.170091","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Spawn loader worker #1 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171740","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Ordinary database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171798","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Ordinary database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171812","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Spawn loader worker #2 in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171815","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171840","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","SOURCE_FILE":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171891","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171899","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Execute load job 'startup Atomic database default' in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171906","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","SOURCE_FILE":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171911","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Finish load job 'startup Atomic database default' with status OK","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171923","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Loaded metadata.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.171932","THREAD_NAME":"","THREAD_ID":"667","LEVEL":"Debug","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172069","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172033","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","MESSAGE":"Stop worker in BackgrndStartup","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172106","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Tasks stats provider: procfs","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172104","THREAD_NAME":"","THREAD_ID":"669","LEVEL":"Debug","MESSAGE":"Change current priority: 2 -> none","SOURCE_FILE":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172190","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"There are 0 detached tables. Start searching non used tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172205","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Debug","MESSAGE":"Found 0 non used tables in detached tables.","SOURCE_FILE":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172492","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172541","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.172549","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Trace","MESSAGE":"Generating new RSA key pair.","SOURCE_FILE":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.226972","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.226997","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","SOURCE_FILE":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230655","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for http:\/\/0.0.0.0:8123","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230720","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for native protocol (tcp): 0.0.0.0:9000","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230771","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230823","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.230845","THREAD_NAME":"","THREAD_ID":"1","LEVEL":"Information","MESSAGE":"Ready for connections.","SOURCE_FILE":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231790","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","MESSAGE":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231896","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"TCP Request. Address: 172.16.1.1:56132","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231977","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Information","MESSAGE":"Client has not sent any data.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.231988","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628542","THREAD_NAME":"TCPServer: 0.0.0.0:9000","THREAD_ID":"675","LEVEL":"Test","MESSAGE":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","SOURCE_FILE":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628643","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"TCP Request. Address: 172.16.1.1:56136","SOURCE_FILE":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628738","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628764","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Authenticating user 'default' from 172.16.1.1:56136","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628811","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628837","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.628971","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.629035","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.629069","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","SOURCE_FILE":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638472","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","SOURCE_FILE":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638829","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"(from 172.16.1.1:56136) (query 1, line 1) SELECT 1 (stage: Complete)","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.638875","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Did not add the task because the timeout is 0. Query: SELECT 1","SOURCE_FILE":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639064","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Query to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639145","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Trace","MESSAGE":"Query from stage FetchColumns to stage Complete","SOURCE_FILE":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.639280","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Scheduling next collecting task for query_id 0047e1e3-8089-4426-996a-de12bf1f6cf9 in 999 ms","SOURCE_FILE":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.640101","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Read 1 rows, 1.00 B in 0.001285 sec., 778.2101167315176 rows\/sec., 778.21 B\/sec.","SOURCE_FILE":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.640237","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Processed in 0.001937624 sec.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650021","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Test","MESSAGE":"Closing connection (open: true, cancelled: false, eof: true)","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650047","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"Done processing connection.","SOURCE_FILE":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()"} {"DATE_TIME_UTC":"2025-04-02T04:15:43Z","DATE_TIME":"1743567343.650070","THREAD_NAME":"TCPServerConnection ([#1])","THREAD_ID":"10","LEVEL":"Debug","MESSAGE":"170768fd-8c42-4cad-85c8-191eac98fd39 Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","SOURCE_FILE":"src\/Interpreters\/Session.cpp; DB::Session::~Session()"} Executing query SELECT 1 on node_no_keys Executing query SELECT 1 on node_no_keys Executing query select count() from system.parts where table = 'zero_copy_mutation' and active on node2 Executing query select count() from system.parts where table = 'zero_copy_mutation' and active on node2 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (87, '0x1d91') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (87, '0x1d91') on node grep in log called grep in log called run container_id:rootteststructuredloggingjson-gw8-node_no_keys-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true'] run container_id:rootteststructuredloggingjson-gw8-node_no_keys-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true'] Command:[docker exec rootteststructuredloggingjson-gw8-node_no_keys-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true] Command:[docker exec rootteststructuredloggingjson-gw8-node_no_keys-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -a "" /var/log/clickhouse-server/clickhouse-server.log* || true] Stderr: Container roottests3tablefunctions-gw5-node-1 Stopping Stderr: Container roottests3tablefunctions-gw5-node-1 Stopping Stderr: Container roottests3tablefunctions-gw5-resolver-1 Stopping Stderr: Container roottests3tablefunctions-gw5-resolver-1 Stopping Stderr: Container roottests3tablefunctions-gw5-node-1 Stopped Stderr: Container roottests3tablefunctions-gw5-node-1 Stopped Stderr: Container roottests3tablefunctions-gw5-node-1 Removing Stderr: Container roottests3tablefunctions-gw5-node-1 Removing Stderr: Container roottests3tablefunctions-gw5-resolver-1 Stopped Stderr: Container roottests3tablefunctions-gw5-resolver-1 Stopped Stderr: Container roottests3tablefunctions-gw5-resolver-1 Removing Stderr: Container roottests3tablefunctions-gw5-resolver-1 Removing Stderr: Container roottests3tablefunctions-gw5-node-1 Removed Stderr: Container roottests3tablefunctions-gw5-node-1 Removed Stderr: Container roottests3tablefunctions-gw5-minio1-1 Stopping Stderr: Container roottests3tablefunctions-gw5-minio1-1 Stopping Stderr: Container roottests3tablefunctions-gw5-minio1-1 Stopped Stderr: Container roottests3tablefunctions-gw5-minio1-1 Stopped Stderr: Container roottests3tablefunctions-gw5-minio1-1 Removing Stderr: Container roottests3tablefunctions-gw5-minio1-1 Removing Stderr: Container roottests3tablefunctions-gw5-resolver-1 Removed Stderr: Container roottests3tablefunctions-gw5-resolver-1 Removed Stderr: Container roottests3tablefunctions-gw5-minio1-1 Removed Stderr: Container roottests3tablefunctions-gw5-minio1-1 Removed Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Stopping Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Stopping Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Stopping Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Stopping Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Stopped Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Stopped Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Removing Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Removing Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Stopped Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Stopped Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Removing Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Removing Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Removed Stderr: Container roottests3tablefunctions-gw5-proxy2-1 Removed Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Removed Stderr: Container roottests3tablefunctions-gw5-proxy1-1 Removed Stderr: Volume roottests3tablefunctions-gw5_data1-1 Removing Stderr: Volume roottests3tablefunctions-gw5_data1-1 Removing Stderr: Network roottests3tablefunctions-gw5_default Removing Stderr: Network roottests3tablefunctions-gw5_default Removing Stderr: Volume roottests3tablefunctions-gw5_data1-1 Removed Stderr: Volume roottests3tablefunctions-gw5_data1-1 Removed Stderr: Network roottests3tablefunctions-gw5_default Removed Stderr: Network roottests3tablefunctions-gw5_default Removed Cleanup called Cleanup called Executing query SELECT arrayJoin(data_paths) FROM system.tables WHERE name='zero_copy_mutation' on node1 Executing query SELECT arrayJoin(data_paths) FROM system.tables WHERE name='zero_copy_mutation' on node1 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (88, '0x1e40') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (88, '0x1e40') on node Docker networks for project roottests3tablefunctions-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3tablefunctions-gw5 are NETWORK ID NAME DRIVER SCOPE Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.814237","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"SentryWriter","message":"Sending crash reports is disabled","source_file":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)","source_line":"144"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.814237","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"SentryWriter","message":"Sending crash reports is disabled","source_file":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)","source_line":"144"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873243","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","source_file":"","source_line":"0"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873243","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","source_file":"","source_line":"0"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873394","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"starting up","source_file":"","source_line":"0"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873394","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"starting up","source_file":"","source_line":"0"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873406","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","source_file":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)","source_line":"585"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873406","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","source_file":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)","source_line":"585"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873517","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Jemalloc","message":"Value for background_thread set to true (from true)","source_file":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]","source_line":"32"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873517","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Jemalloc","message":"Value for background_thread set to true (from true)","source_file":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]","source_line":"32"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877381","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1027"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877381","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1027"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877411","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1042"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877411","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1042"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877439","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Pipe","message":"Pipe capacity is 1.00 MiB","source_file":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)","source_line":"131"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877439","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Pipe","message":"Pipe capacity is 1.00 MiB","source_file":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)","source_line":"131"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878190","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CgroupsReader","message":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","source_file":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)","source_line":"214"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878190","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CgroupsReader","message":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","source_file":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)","source_line":"214"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878346","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/class\/thermal","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()","source_line":"126"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878346","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/class\/thermal","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()","source_line":"126"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878367","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/block","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()","source_line":"163"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878367","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/block","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()","source_line":"163"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878458","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/devices\/system\/edac","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()","source_line":"190"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878458","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/devices\/system\/edac","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()","source_line":"190"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878481","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/class\/hwmon","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()","source_line":"220"} Stdout:{"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878481","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/class\/hwmon","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()","source_line":"220"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.030283","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1349"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.030283","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1349"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.030413","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1403"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.030413","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1403"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034661","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1407"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034661","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1407"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034707","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"MemoryWorker","message":"Starting background memory thread with period of 50ms, using Cgroups as source","source_file":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()","source_line":"252"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034707","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"MemoryWorker","message":"Starting background memory thread with period of 50ms, using Cgroups as source","source_file":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()","source_line":"252"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034786","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgSchPool","message":"Create BackgroundSchedulePool with 512 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034786","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgSchPool","message":"Create BackgroundSchedulePool with 512 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084290","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"rlimit on number of file descriptors is 1048576","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1474"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084290","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"rlimit on number of file descriptors is 1048576","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1474"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084335","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"rlimit on number of threads is 18446744073709551615","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1496"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084335","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"rlimit on number of threads is 18446744073709551615","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1496"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084355","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Initializing DateLUT.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1524"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084355","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Initializing DateLUT.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1524"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084362","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"Initialized DateLUT with time zone 'Etc\/UTC'.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1526"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084362","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"Initialized DateLUT with time zone 'Etc\/UTC'.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1526"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084341","thread_name":"","thread_id":"16","level":"Trace","query_id":"","logger_name":"CancellationChecker","message":"Started worker function","source_file":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()","source_line":"97"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084341","thread_name":"","thread_id":"16","level":"Trace","query_id":"","logger_name":"CancellationChecker","message":"Started worker function","source_file":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()","source_line":"97"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084412","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Context","message":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","source_file":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)","source_line":"1324"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084412","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Context","message":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","source_file":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)","source_line":"1324"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084827","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_no_keys' as replica host.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1605"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084827","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_no_keys' as replica host.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1605"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084847","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Initializing interserver credentials.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1619"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084847","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Initializing interserver credentials.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1619"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084948","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"NamedCollectionsMetadataStorage","message":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","source_file":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)","source_line":"627"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084948","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"NamedCollectionsMetadataStorage","message":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","source_file":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)","source_line":"627"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084968","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from config","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)","source_line":"256"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084968","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from config","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)","source_line":"256"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084981","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from sql","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)","source_line":"276"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084981","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from sql","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)","source_line":"276"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084992","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"FileCacheFactory","message":"Will load 0 caches from default cache config","source_file":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)","source_line":"226"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084992","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"FileCacheFactory","message":"Will load 0 caches from default cache config","source_file":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)","source_line":"226"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.087086","thread_name":"","thread_id":"15","level":"Information","query_id":"","logger_name":"MemoryTracker","message":"Correcting the value of global memory tracker from 11.08 MiB to 123.43 MiB","source_file":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)","source_line":"542"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.087086","thread_name":"","thread_id":"15","level":"Information","query_id":"","logger_name":"MemoryTracker","message":"Correcting the value of global memory tracker from 11.08 MiB to 123.43 MiB","source_file":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)","source_line":"542"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.087987","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loading config '\/etc\/clickhouse-server\/config.xml'","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"128"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.087987","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loading config '\/etc\/clickhouse-server\/config.xml'","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"128"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088008","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"680"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088008","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"680"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088849","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088849","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088940","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088940","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.089009","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.089009","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.089170","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_no_keys_json.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.089170","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_no_keys_json.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.093222","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","source_line":"926"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.093222","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","source_line":"926"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.093330","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"175"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.093330","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"175"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099442","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1792"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099442","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1792"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099461","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1819"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099461","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1819"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099467","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Merges and mutations memory limit is set to 15.30 GiB","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1832"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099467","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Merges and mutations memory limit is set to 15.30 GiB","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1832"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099579","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting max_remote_read_network_bandwidth_for_server was set to 0","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1874"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099579","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting max_remote_read_network_bandwidth_for_server was set to 0","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1874"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099586","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting max_remote_write_network_bandwidth_for_server was set to 0","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1875"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099586","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting max_remote_write_network_bandwidth_for_server was set to 0","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1875"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099596","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1886"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099596","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1886"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099608","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgBufSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099608","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgBufSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.100952","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgMBSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.100952","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgMBSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.102404","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgDistSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.102404","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgDistSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104329","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from config","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)","source_line":"267"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104329","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from config","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)","source_line":"267"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104364","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"193"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104364","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"193"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104388","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ConfigReloader","message":"Config reload interval set to 2000ms","source_file":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","source_line":"45"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104388","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ConfigReloader","message":"Config reload interval set to 2000ms","source_file":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","source_line":"45"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104827","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2187"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104827","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2187"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107143","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loading config '\/etc\/clickhouse-server\/users.xml'","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"128"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107143","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loading config '\/etc\/clickhouse-server\/users.xml'","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"128"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107292","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"680"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107292","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"680"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107498","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107498","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107543","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107543","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107580","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107580","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","source_line":"926"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","source_line":"926"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108041","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"175"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108041","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"175"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108525","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"193"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108525","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"193"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108602","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ConfigReloader","message":"Config reload interval set to 2000ms","source_file":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","source_line":"45"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108602","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ConfigReloader","message":"Config reload interval set to 2000ms","source_file":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","source_line":"45"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108837","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Access(user directories)","message":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","source_file":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)","source_line":"359"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108837","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Access(user directories)","message":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","source_file":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)","source_line":"359"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108952","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Access(local_directory)","message":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","source_file":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()","source_line":"246"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108952","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Access(local_directory)","message":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","source_file":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()","source_line":"246"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108975","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Access(local_directory)","message":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","source_file":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)","source_line":"186"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108975","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Access(local_directory)","message":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","source_file":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)","source_line":"186"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109087","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Access(user directories)","message":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","source_file":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)","source_line":"397"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109087","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Access(user directories)","message":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","source_file":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)","source_line":"397"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109168","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CgroupsMemoryUsageObserver","message":"Started cgroup current memory usage observer thread","source_file":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()","source_line":"43"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109168","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CgroupsMemoryUsageObserver","message":"Started cgroup current memory usage observer thread","source_file":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()","source_line":"43"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109477","thread_name":"","thread_id":"582","level":"Information","query_id":"","logger_name":"CgroupsMemoryUsageObserver","message":"Memory amount initially available to the process is 30.60 GiB","source_file":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()","source_line":"67"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109477","thread_name":"","thread_id":"582","level":"Information","query_id":"","logger_name":"CgroupsMemoryUsageObserver","message":"Memory amount initially available to the process is 30.60 GiB","source_file":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()","source_line":"67"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.110985","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5940"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.110985","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5940"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.111896","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for move operations with num_threads=8, num_tasks=8","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5950"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.111896","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for move operations with num_threads=8, num_tasks=8","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5950"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.113270","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for fetches with num_threads=16, num_tasks=16","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5960"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.113270","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for fetches with num_threads=16, num_tasks=16","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5960"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.114684","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5970"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.114684","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5970"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119071","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Context","message":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","source_file":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","source_line":"932"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119071","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Context","message":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","source_file":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","source_line":"932"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119653","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Context","message":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","source_file":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","source_line":"932"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119653","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Context","message":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","source_file":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","source_line":"932"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119893","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DNSCacheUpdater","message":"Update period 15 seconds","source_file":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()","source_line":"49"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119893","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DNSCacheUpdater","message":"Update period 15 seconds","source_file":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()","source_line":"49"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119913","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Loading metadata from \/var\/lib\/clickhouse\/","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2315"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119913","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Loading metadata from \/var\/lib\/clickhouse\/","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2315"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119951","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Database disk name: default","source_file":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","source_line":"1182"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119951","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Database disk name: default","source_file":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","source_line":"1182"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119958","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Database disk name: default, path: \/var\/lib\/clickhouse\/","source_file":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","source_line":"1190"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119958","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Database disk name: default, path: \/var\/lib\/clickhouse\/","source_file":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","source_line":"1190"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119969","thread_name":"","thread_id":"17","level":"Debug","query_id":"","logger_name":"DNSResolver","message":"Updating DNS cache","source_file":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","source_line":"446"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119969","thread_name":"","thread_id":"17","level":"Debug","query_id":"","logger_name":"DNSResolver","message":"Updating DNS cache","source_file":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","source_line":"446"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.120065","thread_name":"","thread_id":"17","level":"Debug","query_id":"","logger_name":"DNSResolver","message":"Updated DNS cache","source_file":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","source_line":"477"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.120065","thread_name":"","thread_id":"17","level":"Debug","query_id":"","logger_name":"DNSResolver","message":"Updated DNS cache","source_file":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","source_line":"477"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.120009","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/system.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.120009","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/system.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.123722","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/information_schema.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.123722","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/information_schema.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.125439","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.125439","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126783","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","source_file":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","source_line":"311"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126783","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","source_file":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","source_line":"311"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126867","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"TablesLoader","message":"Parsed metadata of 0 tables in 1 databases in 6.4481e-05 sec","source_file":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","source_line":"63"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126867","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"TablesLoader","message":"Parsed metadata of 0 tables in 1 databases in 6.4481e-05 sec","source_file":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","source_line":"63"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126892","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ReferentialDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126892","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ReferentialDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126902","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"LoadingDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126902","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"LoadingDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126949","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","source_line":"693"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126949","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","source_line":"693"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126995","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","source_line":"693"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126995","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","source_line":"693"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127019","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127019","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127029","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Atomic database system' into ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127029","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Atomic database system' into ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127040","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: none -> 0","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127040","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: none -> 0","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127049","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #1 in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127049","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #1 in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127173","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Wait load job 'startup Atomic database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)","source_line":"802"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127173","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Wait load job 'startup Atomic database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)","source_line":"802"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127250","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Ordinary database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127250","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Ordinary database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127325","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Ordinary database system' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127325","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Ordinary database system' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127350","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #2 in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127350","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #2 in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127685","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Atomic database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} Docker containers for project roottests3tablefunctions-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127685","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Atomic database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} Docker containers for project roottests3tablefunctions-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127724","thread_name":"","thread_id":"648","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127724","thread_name":"","thread_id":"648","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127818","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Atomic database system' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127818","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Atomic database system' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127830","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127830","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127837","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: 0 -> none","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127837","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: 0 -> none","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127862","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_log from query_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127862","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_log from query_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.129919","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_thread_log from query_thread_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.129919","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_thread_log from query_thread_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.130583","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.part_log from part_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.130583","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.part_log from part_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.131104","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.trace_log from trace_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.131104","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.trace_log from trace_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.131563","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.crash_log from crash_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.131563","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.crash_log from crash_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.132503","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.text_log from text_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.132503","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.text_log from text_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.133105","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.metric_log from metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.133105","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.metric_log from metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.140638","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.latency_log from latency_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.140638","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.latency_log from latency_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.140928","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.error_log from error_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.140928","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.error_log from error_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141147","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141147","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141221","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141221","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141242","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.s3queue_log from s3queue_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141242","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.s3queue_log from s3queue_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141497","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141497","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141516","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.asynchronous_metric_log from asynchronous_metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141516","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.asynchronous_metric_log from asynchronous_metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141715","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.opentelemetry_span_log from opentelemetry_span_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141715","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.opentelemetry_span_log from opentelemetry_span_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141990","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_views_log from query_views_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141990","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_views_log from query_views_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142345","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142345","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142372","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.session_log since corresponding section 'session_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142372","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.session_log since corresponding section 'session_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142385","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142385","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]","source_line":"153"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142397","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.processors_profile_log from processors_profile_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142397","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.processors_profile_log from processors_profile_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142770","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.asynchronous_insert_log from asynchronous_insert_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142770","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.asynchronous_insert_log from asynchronous_insert_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143132","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.backup_log from backup_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143132","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.backup_log from backup_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143443","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.blob_storage_log from blob_storage_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143443","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.blob_storage_log from blob_storage_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143675","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_metric_log from query_metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143675","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_metric_log from query_metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]","source_line":"158"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.149124","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.149124","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.149151","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.149151","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150446","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150446","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150457","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150457","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150483","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150483","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150489","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150489","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150508","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150508","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150516","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150516","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150535","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150535","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150540","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150540","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150582","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150582","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150593","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150593","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150693","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150693","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150700","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150700","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150750","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150750","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150756","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150756","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150813","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150813","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150821","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150821","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150866","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150866","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150872","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150872","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150907","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150907","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150913","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150913","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150957","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150957","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150963","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150963","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150997","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150997","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151003","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151003","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151041","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151041","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151047","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151047","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151088","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151088","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151095","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151095","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151137","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151137","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151143","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151143","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151168","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151168","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151178","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151178","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151202","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151202","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151207","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151207","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151230","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151230","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151272","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151272","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151323","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151323","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151330","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151330","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151353","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151353","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151358","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151358","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151390","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151390","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151400","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151400","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151480","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151480","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151488","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151488","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151524","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151524","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151569","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151569","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151604","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151604","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151609","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151609","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151681","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151681","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151687","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151687","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151743","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151743","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151750","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151750","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151919","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151919","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151927","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151927","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151959","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151959","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151965","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151965","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151989","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151989","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151995","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151995","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152021","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152021","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152057","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152057","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152064","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152064","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152118","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152118","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152124","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152124","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152163","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152163","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152169","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152169","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152221","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152221","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152226","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152226","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152280","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152280","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152286","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152286","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152422","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152422","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152431","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152431","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152563","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152563","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152578","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152578","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152628","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152628","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152634","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152634","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152760","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152760","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152767","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152767","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152796","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152796","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152803","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152803","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152823","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152823","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152828","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152828","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152852","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152852","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152863","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152863","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152937","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152937","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152945","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152945","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152971","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152971","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152977","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152977","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152996","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152996","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153000","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153000","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153048","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153048","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153054","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153054","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153092","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153092","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153099","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153099","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153134","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153134","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153141","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153141","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153331","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153331","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153339","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153339","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153437","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153437","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153444","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153444","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153481","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153481","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153486","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153486","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153505","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153505","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153512","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153512","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153555","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153555","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153564","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153564","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153611","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153611","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153617","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153617","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153640","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153640","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153645","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153645","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153665","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153665","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153669","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153669","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153851","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153851","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153862","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153862","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154005","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154005","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154017","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154017","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154080","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154080","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154086","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154086","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154253","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154253","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154262","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154262","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154381","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154381","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154388","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154388","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154438","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154438","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154445","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154445","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154491","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154491","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154496","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154496","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154663","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154663","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154671","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154671","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154729","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154729","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154735","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154735","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154784","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154784","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154790","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154790","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154824","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154824","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154830","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154830","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154872","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154872","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Docker volumes for project roottests3tablefunctions-gw5 are DRIVER VOLUME NAME Docker volumes for project roottests3tablefunctions-gw5 are DRIVER VOLUME NAME Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154878","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154878","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Command:[docker container list --all --filter name='^/roottests3tablefunctions-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottests3tablefunctions-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155027","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155027","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155085","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155085","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155091","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155091","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155149","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155149","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155156","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155156","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155193","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155193","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155199","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155199","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155264","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155264","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155270","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155270","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155319","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155319","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155329","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155329","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155387","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155387","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155394","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155394","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155458","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155458","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155466","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155466","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155491","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155491","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155497","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155497","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155528","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155528","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155533","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155533","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155573","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155573","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155583","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155583","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155619","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155619","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155625","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155625","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155660","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155660","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155694","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155694","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155778","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155778","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155785","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155785","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155826","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155826","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155832","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155832","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155859","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155859","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155864","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155864","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155907","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155907","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155912","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155912","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155941","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155941","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155946","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155946","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156005","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156005","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156011","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156011","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156089","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156089","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156096","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156096","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156123","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156123","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156128","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156128","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156163","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156163","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156169","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156169","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156201","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156201","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156207","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156207","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156235","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156235","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156240","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156240","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156269","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156269","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156274","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156274","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156295","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156295","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156314","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156314","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156363","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156363","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156373","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156373","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156404","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156404","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156409","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156409","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156435","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156435","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156440","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156440","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156521","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156521","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156527","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156527","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.163703","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseCatalog","message":"Found 0 partially dropped tables. Will load them and retry removal.","source_file":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()","source_line":"1050"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.163703","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseCatalog","message":"Found 0 partially dropped tables. Will load them and retry removal.","source_file":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()","source_line":"1050"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165325","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseAtomic (default)","message":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","source_file":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","source_line":"311"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165325","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseAtomic (default)","message":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","source_file":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","source_line":"311"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165356","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"TablesLoader","message":"Parsed metadata of 0 tables in 1 databases in 9.2864e-05 sec","source_file":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","source_line":"63"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165356","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"TablesLoader","message":"Parsed metadata of 0 tables in 1 databases in 9.2864e-05 sec","source_file":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","source_line":"63"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165381","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ReferentialDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165381","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ReferentialDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165389","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"LoadingDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165389","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"LoadingDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165432","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"loadMetadata","message":"Start asynchronous loading of databases","source_file":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)","source_line":"268"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165432","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"loadMetadata","message":"Start asynchronous loading of databases","source_file":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)","source_line":"268"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165454","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165454","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165463","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Atomic database default' into BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165463","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Atomic database default' into BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165474","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: none -> 2","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165474","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: none -> 2","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165482","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #1 in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165482","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #1 in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165670","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"UserDefinedSQLObjectsLoaderFromDisk","message":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","source_file":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","source_line":"131"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165670","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"UserDefinedSQLObjectsLoaderFromDisk","message":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","source_file":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","source_line":"131"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165691","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"UserDefinedSQLObjectsLoaderFromDisk","message":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","source_file":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","source_line":"135"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165691","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"UserDefinedSQLObjectsLoaderFromDisk","message":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","source_file":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","source_line":"135"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165690","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Ordinary database default' in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165690","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Ordinary database default' in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165741","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Ordinary database default' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165741","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Ordinary database default' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165748","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"WorkloadEntityDiskStorage","message":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","source_file":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","source_line":"129"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165748","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"WorkloadEntityDiskStorage","message":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","source_file":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","source_line":"129"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165762","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #2 in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165762","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #2 in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165787","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"WorkloadEntityDiskStorage","message":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","source_file":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","source_line":"133"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165787","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"WorkloadEntityDiskStorage","message":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","source_file":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","source_line":"133"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165801","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Loaded metadata.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2408"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165801","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Loaded metadata.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2408"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165876","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Atomic database default' in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165876","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Atomic database default' in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165887","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Atomic database default' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165887","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Atomic database default' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165893","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165893","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165903","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165903","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165915","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: 2 -> none","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165915","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: 2 -> none","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165957","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"bool DB::(anonymous namespace)::checkPermissionsImpl()","message":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()","source_line":"216"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165957","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"bool DB::(anonymous namespace)::checkPermissionsImpl()","message":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()","source_line":"216"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165987","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Tasks stats provider: procfs","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2428"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165987","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Tasks stats provider: procfs","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2428"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166076","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166076","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166089","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166089","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166466","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","source_line":"37"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166466","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","source_line":"37"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166561","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","source_line":"48"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166561","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","source_line":"48"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166575","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Generating new RSA key pair.","source_file":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()","source_line":"107"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166575","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Generating new RSA key pair.","source_file":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()","source_line":"107"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.222054","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CertificateReloader","message":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","source_file":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","source_line":"142"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.222054","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CertificateReloader","message":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","source_file":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","source_line":"142"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.222082","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CertificateReloader","message":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","source_file":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","source_line":"142"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.222082","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CertificateReloader","message":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","source_file":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","source_line":"142"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225288","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for http:\/\/0.0.0.0:8123","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225288","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for http:\/\/0.0.0.0:8123","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225355","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for native protocol (tcp): 0.0.0.0:9000","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225355","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for native protocol (tcp): 0.0.0.0:9000","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225405","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225405","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225460","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225460","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225467","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Ready for connections.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2548"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225467","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Ready for connections.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2548"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246774","thread_name":"TCPServer: 0.0.0.0:9000","thread_id":"675","level":"Test","query_id":"","logger_name":"Poco","message":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","source_file":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","source_line":"49"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246774","thread_name":"TCPServer: 0.0.0.0:9000","thread_id":"675","level":"Test","query_id":"","logger_name":"Poco","message":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","source_file":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","source_line":"49"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246896","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCPHandlerFactory","message":"TCP Request. Address: 172.16.1.1:64780","source_file":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","source_line":"54"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246896","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCPHandlerFactory","message":"TCP Request. Address: 172.16.1.1:64780","source_file":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","source_line":"54"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246975","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Information","query_id":"","logger_name":"TCPHandler","message":"Client has not sent any data.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","source_line":"330"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246975","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Information","query_id":"","logger_name":"TCPHandler","message":"Client has not sent any data.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","source_line":"330"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246986","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Done processing connection.","source_file":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","source_line":"2630"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246986","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Done processing connection.","source_file":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","source_line":"2630"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774050","thread_name":"TCPServer: 0.0.0.0:9000","thread_id":"675","level":"Test","query_id":"","logger_name":"Poco","message":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","source_file":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","source_line":"49"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774050","thread_name":"TCPServer: 0.0.0.0:9000","thread_id":"675","level":"Test","query_id":"","logger_name":"Poco","message":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","source_file":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","source_line":"49"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774109","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCPHandlerFactory","message":"TCP Request. Address: 172.16.1.1:64794","source_file":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","source_line":"54"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774109","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCPHandlerFactory","message":"TCP Request. Address: 172.16.1.1:64794","source_file":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","source_line":"54"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774175","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()","source_line":"1695"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774175","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()","source_line":"1695"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774206","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"Authenticating user 'default' from 172.16.1.1:64794","source_file":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","source_line":"364"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774206","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"Authenticating user 'default' from 172.16.1.1:64794","source_file":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","source_line":"364"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774269","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","source_line":"374"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774269","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","source_line":"374"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774291","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()","source_line":"534"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774291","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()","source_line":"534"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774407","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"432"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774407","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"432"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774440","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"433"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774440","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"433"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774461","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"434"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774461","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"434"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784574","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","source_file":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const","source_line":"664"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784574","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","source_file":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const","source_line":"664"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784865","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"executeQuery","message":"(from 172.16.1.1:64794) (query 1, line 1) SELECT 1 (stage: Complete)","source_file":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)","source_line":"237"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784865","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"executeQuery","message":"(from 172.16.1.1:64794) (query 1, line 1) SELECT 1 (stage: Complete)","source_file":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)","source_line":"237"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784896","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"CancellationChecker","message":"Did not add the task because the timeout is 0. Query: SELECT 1","source_file":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)","source_line":"77"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784896","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"CancellationChecker","message":"Did not add the task because the timeout is 0. Query: SELECT 1","source_file":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)","source_line":"77"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785063","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"Planner","message":"Query to stage Complete","source_file":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()","source_line":"1296"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785063","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"Planner","message":"Query to stage Complete","source_file":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()","source_line":"1296"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785140","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"Planner","message":"Query from stage FetchColumns to stage Complete","source_file":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()","source_line":"1574"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785140","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"Planner","message":"Query from stage FetchColumns to stage Complete","source_file":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()","source_line":"1574"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785260","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"QueryMetricLog","message":"Scheduling next collecting task for query_id 1059a1fa-69ac-4a60-8029-f5334c480773 in 999 ms","source_file":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)","source_line":"219"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785260","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"QueryMetricLog","message":"Scheduling next collecting task for query_id 1059a1fa-69ac-4a60-8029-f5334c480773 in 999 ms","source_file":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)","source_line":"219"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785920","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"executeQuery","message":"Read 1 rows, 1.00 B in 0.001067 sec., 937.207122774133 rows\/sec., 937.21 B\/sec.","source_file":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)","source_line":"592"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785920","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"executeQuery","message":"Read 1 rows, 1.00 B in 0.001067 sec., 937.207122774133 rows\/sec., 937.21 B\/sec.","source_file":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)","source_line":"592"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.786033","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"TCPHandler","message":"Processed in 0.001574242 sec.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)","source_line":"865"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.786033","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"TCPHandler","message":"Processed in 0.001574242 sec.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)","source_line":"865"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795736","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"","logger_name":"TCPHandler","message":"Closing connection (open: true, cancelled: false, eof: true)","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","source_line":"456"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795736","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"","logger_name":"TCPHandler","message":"Closing connection (open: true, cancelled: false, eof: true)","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","source_line":"456"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795753","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Done processing connection.","source_file":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","source_line":"2630"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795753","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Done processing connection.","source_file":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","source_line":"2630"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795778","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; DB::Session::~Session()","source_line":"308"} Stdout:{"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795778","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; DB::Session::~Session()","source_line":"308"} grep result {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.814237","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"SentryWriter","message":"Sending crash reports is disabled","source_file":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)","source_line":"144"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873243","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","source_file":"","source_line":"0"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873394","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"starting up","source_file":"","source_line":"0"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873406","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","source_file":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)","source_line":"585"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873517","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Jemalloc","message":"Value for background_thread set to true (from true)","source_file":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]","source_line":"32"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877381","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1027"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877411","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1042"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877439","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Pipe","message":"Pipe capacity is 1.00 MiB","source_file":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)","source_line":"131"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878190","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CgroupsReader","message":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","source_file":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)","source_line":"214"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878346","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/class\/thermal","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()","source_line":"126"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878367","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/block","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()","source_line":"163"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878458","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/devices\/system\/edac","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()","source_line":"190"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878481","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/class\/hwmon","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()","source_line":"220"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.030283","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1349"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.030413","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1403"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034661","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1407"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034707","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"MemoryWorker","message":"Starting background memory thread with period of 50ms, using Cgroups as source","source_file":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()","source_line":"252"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034786","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgSchPool","message":"Create BackgroundSchedulePool with 512 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084290","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"rlimit on number of file descriptors is 1048576","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1474"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084335","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"rlimit on number of threads is 18446744073709551615","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1496"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084355","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Initializing DateLUT.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1524"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084362","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"Initialized DateLUT with time zone 'Etc\/UTC'.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1526"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084341","thread_name":"","thread_id":"16","level":"Trace","query_id":"","logger_name":"CancellationChecker","message":"Started worker function","source_file":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()","source_line":"97"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084412","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Context","message":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","source_file":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)","source_line":"1324"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084827","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_no_keys' as replica host.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1605"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084847","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Initializing interserver credentials.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1619"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084948","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"NamedCollectionsMetadataStorage","message":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","source_file":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)","source_line":"627"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084968","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from config","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)","source_line":"256"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084981","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from sql","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)","source_line":"276"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084992","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"FileCacheFactory","message":"Will load 0 caches from default cache config","source_file":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)","source_line":"226"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.087086","thread_name":"","thread_id":"15","level":"Information","query_id":"","logger_name":"MemoryTracker","message":"Correcting the value of global memory tracker from 11.08 MiB to 123.43 MiB","source_file":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)","source_line":"542"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.087987","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loading config '\/etc\/clickhouse-server\/config.xml'","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"128"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088008","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"680"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088849","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088940","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.089009","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.089170","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_no_keys_json.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.093222","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","source_line":"926"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.093330","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"175"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099442","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1792"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099461","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1819"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099467","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Merges and mutations memory limit is set to 15.30 GiB","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1832"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099579","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting max_remote_read_network_bandwidth_for_server was set to 0","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1874"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099586","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting max_remote_write_network_bandwidth_for_server was set to 0","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1875"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099596","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1886"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099608","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgBufSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.100952","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgMBSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.102404","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgDistSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104329","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from config","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)","source_line":"267"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104364","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"193"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104388","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ConfigReloader","message":"Config reload interval set to 2000ms","source_file":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","source_line":"45"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104827","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2187"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107143","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loading config '\/etc\/clickhouse-server\/users.xml'","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"128"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107292","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"680"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107498","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107543","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107580","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","source_line":"926"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108041","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"175"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108525","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"193"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108602","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ConfigReloader","message":"Config reload interval set to 2000ms","source_file":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","source_line":"45"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108837","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Access(user directories)","message":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","source_file":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)","source_line":"359"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108952","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Access(local_directory)","message":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","source_file":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()","source_line":"246"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108975","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Access(local_directory)","message":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","source_file":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)","source_line":"186"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109087","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Access(user directories)","message":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","source_file":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)","source_line":"397"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109168","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CgroupsMemoryUsageObserver","message":"Started cgroup current memory usage observer thread","source_file":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()","source_line":"43"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109477","thread_name":"","thread_id":"582","level":"Information","query_id":"","logger_name":"CgroupsMemoryUsageObserver","message":"Memory amount initially available to the process is 30.60 GiB","source_file":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()","source_line":"67"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.110985","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5940"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.111896","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for move operations with num_threads=8, num_tasks=8","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5950"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.113270","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for fetches with num_threads=16, num_tasks=16","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5960"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.114684","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5970"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119071","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Context","message":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","source_file":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","source_line":"932"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119653","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Context","message":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","source_file":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","source_line":"932"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119893","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DNSCacheUpdater","message":"Update period 15 seconds","source_file":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()","source_line":"49"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119913","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Loading metadata from \/var\/lib\/clickhouse\/","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2315"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119951","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Database disk name: default","source_file":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","source_line":"1182"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119958","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Database disk name: default, path: \/var\/lib\/clickhouse\/","source_file":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","source_line":"1190"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119969","thread_name":"","thread_id":"17","level":"Debug","query_id":"","logger_name":"DNSResolver","message":"Updating DNS cache","source_file":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","source_line":"446"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.120065","thread_name":"","thread_id":"17","level":"Debug","query_id":"","logger_name":"DNSResolver","message":"Updated DNS cache","source_file":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","source_line":"477"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.120009","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/system.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.123722","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/information_schema.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.125439","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126783","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","source_file":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","source_line":"311"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126867","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"TablesLoader","message":"Parsed metadata of 0 tables in 1 databases in 6.4481e-05 sec","source_file":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","source_line":"63"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126892","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ReferentialDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126902","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"LoadingDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126949","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","source_line":"693"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126995","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","source_line":"693"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127019","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127029","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Atomic database system' into ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127040","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: none -> 0","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127049","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #1 in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127173","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Wait load job 'startup Atomic database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)","source_line":"802"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127250","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Ordinary database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127325","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Ordinary database system' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127350","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #2 in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127685","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Atomic database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127724","thread_name":"","thread_id":"648","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127818","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Atomic database system' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127830","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127837","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: 0 -> none","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127862","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_log from query_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.129919","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_thread_log from query_thread_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.130583","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.part_log from part_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.131104","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.trace_log from trace_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.131563","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.crash_log from crash_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.132503","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.text_log from text_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.133105","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.metric_log from metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.140638","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.latency_log from latency_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.140928","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.error_log from error_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141147","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141221","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141242","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.s3queue_log from s3queue_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141497","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141516","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.asynchronous_metric_log from asynchronous_metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141715","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.opentelemetry_span_log from opentelemetry_span_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141990","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_views_log from query_views_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142345","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142372","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.session_log since corresponding section 'session_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142385","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142397","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.processors_profile_log from processors_profile_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142770","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.asynchronous_insert_log from asynchronous_insert_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143132","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.backup_log from backup_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143443","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.blob_storage_log from blob_storage_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143675","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_metric_log from query_metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.149124","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.149151","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150446","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150457","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150483","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150489","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150508","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150516","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150535","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150540","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150582","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150593","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150693","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150700","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150750","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150756","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150813","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150821","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150866","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150872","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150907","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150913","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150957","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150963","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150997","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151003","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151041","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151047","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151088","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151095","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151137","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151143","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151168","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151178","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151202","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151207","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151230","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151272","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151323","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151330","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151353","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151358","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151390","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151400","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151480","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151488","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151524","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151569","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151604","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151609","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151681","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151687","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151743","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151750","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151919","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151927","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151959","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151965","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151989","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151995","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152021","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152057","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152064","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152118","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152124","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152163","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152169","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152221","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152226","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152280","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152286","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152422","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152431","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152563","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152578","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152628","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152634","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152760","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152767","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152796","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152803","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152823","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152828","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152852","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152863","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152937","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152945","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152971","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152977","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152996","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153000","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153048","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153054","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153092","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153099","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153134","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153141","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153331","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153339","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153437","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153444","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153481","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153486","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153505","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153512","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153555","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153564","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153611","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153617","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153640","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153645","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153665","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153669","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153851","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153862","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154005","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154017","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154080","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154086","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154253","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154262","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154381","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154388","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154438","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154445","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154491","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154496","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154663","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154671","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154729","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154735","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154784","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154790","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154824","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154830","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154872","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154878","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155027","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155085","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155091","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155149","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155156","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155193","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155199","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155264","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155270","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155319","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155329","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155387","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155394","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155458","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155466","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155491","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155497","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155528","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155533","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155573","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155583","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155619","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155625","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155660","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155694","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155778","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155785","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155826","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155832","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155859","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155864","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155907","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155912","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155941","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155946","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156005","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156011","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156089","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156096","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156123","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156128","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156163","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156169","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156201","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156207","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156235","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156240","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156269","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156274","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156295","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156314","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156363","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156373","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156404","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156409","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156435","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156440","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156521","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156527","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.163703","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseCatalog","message":"Found 0 partially dropped tables. Will load them and retry removal.","source_file":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()","source_line":"1050"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165325","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseAtomic (default)","message":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","source_file":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","source_line":"311"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165356","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"TablesLoader","message":"Parsed metadata of 0 tables in 1 databases in 9.2864e-05 sec","source_file":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","source_line":"63"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165381","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ReferentialDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165389","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"LoadingDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165432","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"loadMetadata","message":"Start asynchronous loading of databases","source_file":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)","source_line":"268"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165454","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165463","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Atomic database default' into BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165474","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: none -> 2","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165482","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #1 in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165670","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"UserDefinedSQLObjectsLoaderFromDisk","message":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","source_file":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","source_line":"131"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165691","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"UserDefinedSQLObjectsLoaderFromDisk","message":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","source_file":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","source_line":"135"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165690","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Ordinary database default' in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165741","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Ordinary database default' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165748","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"WorkloadEntityDiskStorage","message":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","source_file":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","source_line":"129"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165762","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #2 in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165787","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"WorkloadEntityDiskStorage","message":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","source_file":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","source_line":"133"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165801","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Loaded metadata.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2408"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165876","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Atomic database default' in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165887","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Atomic database default' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165893","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165903","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165915","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: 2 -> none","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165957","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"bool DB::(anonymous namespace)::checkPermissionsImpl()","message":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()","source_line":"216"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165987","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Tasks stats provider: procfs","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2428"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166076","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166089","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166466","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","source_line":"37"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166561","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","source_line":"48"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166575","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Generating new RSA key pair.","source_file":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()","source_line":"107"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.222054","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CertificateReloader","message":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","source_file":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","source_line":"142"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.222082","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CertificateReloader","message":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","source_file":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","source_line":"142"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225288","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for http:\/\/0.0.0.0:8123","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225355","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for native protocol (tcp): 0.0.0.0:9000","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225405","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225460","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225467","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Ready for connections.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2548"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246774","thread_name":"TCPServer: 0.0.0.0:9000","thread_id":"675","level":"Test","query_id":"","logger_name":"Poco","message":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","source_file":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","source_line":"49"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246896","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCPHandlerFactory","message":"TCP Request. Address: 172.16.1.1:64780","source_file":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","source_line":"54"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246975","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Information","query_id":"","logger_name":"TCPHandler","message":"Client has not sent any data.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","source_line":"330"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246986","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Done processing connection.","source_file":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","source_line":"2630"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774050","thread_name":"TCPServer: 0.0.0.0:9000","thread_id":"675","level":"Test","query_id":"","logger_name":"Poco","message":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","source_file":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","source_line":"49"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774109","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCPHandlerFactory","message":"TCP Request. Address: 172.16.1.1:64794","source_file":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","source_line":"54"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774175","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()","source_line":"1695"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774206","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"Authenticating user 'default' from 172.16.1.1:64794","source_file":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","source_line":"364"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774269","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","source_line":"374"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774291","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()","source_line":"534"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774407","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"432"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774440","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"433"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774461","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"434"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784574","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","source_file":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const","source_line":"664"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784865","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"executeQuery","message":"(from 172.16.1.1:64794) (query 1, line 1) SELECT 1 (stage: Complete)","source_file":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)","source_line":"237"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784896","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"CancellationChecker","message":"Did not add the task because the timeout is 0. Query: SELECT 1","source_file":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)","source_line":"77"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785063","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"Planner","message":"Query to stage Complete","source_file":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()","source_line":"1296"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785140","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"Planner","message":"Query from stage FetchColumns to stage Complete","source_file":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()","source_line":"1574"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785260","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"QueryMetricLog","message":"Scheduling next collecting task for query_id 1059a1fa-69ac-4a60-8029-f5334c480773 in 999 ms","source_file":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)","source_line":"219"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785920","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"executeQuery","message":"Read 1 rows, 1.00 B in 0.001067 sec., 937.207122774133 rows\/sec., 937.21 B\/sec.","source_file":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)","source_line":"592"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.786033","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"TCPHandler","message":"Processed in 0.001574242 sec.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)","source_line":"865"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795736","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"","logger_name":"TCPHandler","message":"Closing connection (open: true, cancelled: false, eof: true)","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","source_line":"456"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795753","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Done processing connection.","source_file":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","source_line":"2630"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795778","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; DB::Session::~Session()","source_line":"308"} grep result {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.814237","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"SentryWriter","message":"Sending crash reports is disabled","source_file":"src\/Daemon\/SentryWriter.cpp; SentryWriter::SentryWriter(Poco::Util::LayeredConfiguration &)","source_line":"144"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873243","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Starting ClickHouse 25.2.2.96963.altinityantalya.96963 (revision: 54496, git hash: 25f7c2a766916d92d37aee82d9defa5f14b86ec8, build id: 4783FC52DD13D5DFB0294BDDD711047195FEB5A6), PID 1","source_file":"","source_line":"0"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873394","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"starting up","source_file":"","source_line":"0"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873406","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"OS name: Linux, version: 5.15.0-130-generic, architecture: x86_64","source_file":"programs\/server\/Server.cpp; virtual void DB::Server::initialize(Poco::Util::Application &)","source_line":"585"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.873517","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Jemalloc","message":"Value for background_thread set to true (from true)","source_file":"src\/Common\/Jemalloc.h; void DB::setJemallocValue(const char *, T) [T = bool]","source_line":"32"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877381","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Available RAM: 30.60 GiB; logical cores: 16; used cores: 16.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1027"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877411","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Available CPU instruction sets: SSE, SSE2, SSE3, SSSE3, SSE41, SSE42, F16C, POPCNT, BMI1, BMI2, PCLMUL, AES, AVX, FMA, AVX2, SHA, ADX, RDRAND, RDSEED, RDTSCP, CLFLUSHOPT, CLWB, XSAVE, OSXSAVE","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1042"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.877439","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Pipe","message":"Pipe capacity is 1.00 MiB","source_file":"src\/Common\/PipeFDs.cpp; void DB::LazyPipeFDs::tryIncreaseSize(int)","source_line":"131"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878190","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CgroupsReader","message":"Will create cgroup reader from '\/sys\/fs\/cgroup\/' (cgroups version: v2)","source_file":"src\/Common\/MemoryWorker.cpp; DB::MemoryWorker::MemoryWorker(uint64_t, bool)","source_line":"214"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878346","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/class\/thermal","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensors()","source_line":"126"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878367","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/block","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openBlockDevices()","source_line":"163"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878458","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/devices\/system\/edac","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openEDAC()","source_line":"190"} {"date_time_utc":"2025-04-02T04:15:42Z","date_time":"1743567342.878481","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"AsynchronousMetrics","message":"Scanning \/sys\/class\/hwmon","source_file":"src\/Common\/AsynchronousMetrics.cpp; void DB::AsynchronousMetrics::openSensorsChips()","source_line":"220"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.030283","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Integrity check of the executable successfully passed (checksum: 2D3F861318CFEDFBBB742B17F83CC397)","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1349"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.030413","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"Will do mlock to prevent executable memory from being paged out. It may take a few seconds.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1403"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034661","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"The memory map of clickhouse executable has been mlock'ed, total 287.17 MiB","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1407"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034707","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"MemoryWorker","message":"Starting background memory thread with period of 50ms, using Cgroups as source","source_file":"src\/Common\/MemoryWorker.cpp; void DB::MemoryWorker::start()","source_line":"252"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.034786","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgSchPool","message":"Create BackgroundSchedulePool with 512 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084290","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"rlimit on number of file descriptors is 1048576","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1474"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084335","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"rlimit on number of threads is 18446744073709551615","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1496"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084355","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Initializing DateLUT.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1524"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084362","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"Application","message":"Initialized DateLUT with time zone 'Etc\/UTC'.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1526"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084341","thread_name":"","thread_id":"16","level":"Trace","query_id":"","logger_name":"CancellationChecker","message":"Started worker function","source_file":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::workerFunction()","source_line":"97"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084412","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Context","message":"Setting up \/var\/lib\/clickhouse\/tmp\/ to store temporary data in it","source_file":"src\/Interpreters\/Context.cpp; void DB::setupTmpPath(LoggerPtr, const std::string &)","source_line":"1324"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084827","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Configuration parameter 'interserver_http_host' doesn't exist or exists and empty. Will use 'node_no_keys' as replica host.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1605"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084847","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Initializing interserver credentials.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"1619"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084948","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"NamedCollectionsMetadataStorage","message":"Using local storage for named collections at path: \/var\/lib\/clickhouse\/named_collections","source_file":"src\/Common\/NamedCollections\/NamedCollectionsMetadataStorage.cpp; static std::unique_ptr DB::NamedCollectionsMetadataStorage::create(const ContextPtr &)","source_line":"627"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084968","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from config","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromConfig(const Poco::Util::AbstractConfiguration &, std::lock_guard &)","source_line":"256"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084981","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from sql","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::loadFromSQL(std::lock_guard &)","source_line":"276"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.084992","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"FileCacheFactory","message":"Will load 0 caches from default cache config","source_file":"src\/Interpreters\/Cache\/FileCacheFactory.cpp; void DB::FileCacheFactory::loadDefaultCaches(const Poco::Util::AbstractConfiguration &)","source_line":"226"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.087086","thread_name":"","thread_id":"15","level":"Information","query_id":"","logger_name":"MemoryTracker","message":"Correcting the value of global memory tracker from 11.08 MiB to 123.43 MiB","source_file":"src\/Common\/MemoryTracker.cpp; static void MemoryTracker::updateAllocated(Int64, bool)","source_line":"542"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.087987","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loading config '\/etc\/clickhouse-server\/config.xml'","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"128"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088008","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Processing configuration file '\/etc\/clickhouse-server\/config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"680"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088849","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.088940","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_enable_keeper_async_replication.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.089009","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/0_common_instance_config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.089170","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/config.d\/config_no_keys_json.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.093222","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/config.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","source_line":"926"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.093330","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/config.xml', performing update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"175"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099442","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Changed setting 'max_server_memory_usage' to 27.54 GiB (30.60 GiB available memory * 0.90 max_server_memory_usage_to_ram_ratio)","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1792"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099461","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting merges_mutations_memory_usage_soft_limit was set to 15.30 GiB (30.60 GiB available * 0.50 merges_mutations_memory_usage_to_ram_ratio)","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1819"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099467","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Merges and mutations memory limit is set to 15.30 GiB","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1832"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099579","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting max_remote_read_network_bandwidth_for_server was set to 0","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1874"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099586","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Setting max_remote_write_network_bandwidth_for_server was set to 0","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1875"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099596","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"ConcurrencyControl limit is set to 32 CPU slots with 'fair_round_robin' scheduler","source_file":"programs\/server\/Server.cpp; auto DB::Server::main(const std::vector &)::(anonymous class)::operator()(ConfigurationPtr, bool) const","source_line":"1886"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.099608","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgBufSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.100952","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgMBSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.102404","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"BackgroundSchedulePool\/BgDistSchPool","message":"Create BackgroundSchedulePool with 16 threads","source_file":"src\/Core\/BackgroundSchedulePool.cpp; DB::BackgroundSchedulePool::BackgroundSchedulePool(size_t, CurrentMetrics::Metric, CurrentMetrics::Metric, const char *)","source_line":"164"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104329","thread_name":"","thread_id":"1","level":"Test","query_id":"","logger_name":"NamedCollectionFactory","message":"Loaded 0 collections from config","source_file":"src\/Common\/NamedCollections\/NamedCollectionsFactory.cpp; void DB::NamedCollectionFactory::reloadFromConfig(const Poco::Util::AbstractConfiguration &)","source_line":"267"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104364","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/config.xml', performed update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"193"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104388","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ConfigReloader","message":"Config reload interval set to 2000ms","source_file":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","source_line":"45"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.104827","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for replica communication (interserver): http:\/\/0.0.0.0:9009","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2187"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107143","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loading config '\/etc\/clickhouse-server\/users.xml'","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"128"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107292","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Processing configuration file '\/etc\/clickhouse-server\/users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"680"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107498","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/conf.d\/macros.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107543","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_common_instance_users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.107580","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Merging configuration file '\/etc\/clickhouse-server\/users.d\/0_random_settings.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; XMLDocumentPtr DB::ConfigProcessor::processConfig(bool *, zkutil::ZooKeeperNodeCache *, const zkutil::EventPtr &, bool)","source_line":"709"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigProcessor","message":"Saved preprocessed configuration to '\/var\/lib\/clickhouse\/preprocessed_configs\/users.xml'.","source_file":"src\/Common\/Config\/ConfigProcessor.cpp; void DB::ConfigProcessor::savePreprocessedConfig(LoadedConfig &, std::string)","source_line":"926"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108041","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/users.xml', performing update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"175"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108525","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"ConfigReloader","message":"Loaded config '\/etc\/clickhouse-server\/users.xml', performed update on configuration","source_file":"src\/Common\/Config\/ConfigReloader.cpp; std::optional DB::ConfigReloader::reloadIfNewer(bool, bool, bool, bool)","source_line":"193"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108602","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ConfigReloader","message":"Config reload interval set to 2000ms","source_file":"src\/Common\/Config\/ConfigReloader.cpp; DB::ConfigReloader::ConfigReloader(std::string_view, const std::vector &, const std::string &, zkutil::ZooKeeperNodeCache &&, const zkutil::EventPtr &, Updater &&)","source_line":"45"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108837","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Access(user directories)","message":"Added users_xml access storage 'users_xml', path: \/etc\/clickhouse-server\/users.xml","source_file":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addUsersConfigStorage(const String &, const String &, const String &, const String &, const zkutil::GetZooKeeper &, bool)","source_line":"359"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108952","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Access(local_directory)","message":"File \/var\/lib\/clickhouse\/access\/users.list doesn't exist","source_file":"src\/Access\/DiskAccessStorage.cpp; bool DB::DiskAccessStorage::readLists()","source_line":"246"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.108975","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Access(local_directory)","message":"Recovering lists in directory \/var\/lib\/clickhouse\/access\/","source_file":"src\/Access\/DiskAccessStorage.cpp; DB::DiskAccessStorage::DiskAccessStorage(const String &, const String &, AccessChangesNotifier &, bool, bool)","source_line":"186"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109087","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Access(user directories)","message":"Added local_directory access storage 'local_directory', path: \/var\/lib\/clickhouse\/access\/","source_file":"src\/Access\/AccessControl.cpp; void DB::AccessControl::addDiskStorage(const String &, const String &, bool, bool)","source_line":"397"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109168","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CgroupsMemoryUsageObserver","message":"Started cgroup current memory usage observer thread","source_file":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::startThread()","source_line":"43"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.109477","thread_name":"","thread_id":"582","level":"Information","query_id":"","logger_name":"CgroupsMemoryUsageObserver","message":"Memory amount initially available to the process is 30.60 GiB","source_file":"src\/Common\/CgroupsMemoryUsageObserver.cpp; void DB::CgroupsMemoryUsageObserver::runThread()","source_line":"67"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.110985","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for merges and mutations with num_threads=16, num_tasks=32, scheduling_policy=round_robin","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5940"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.111896","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for move operations with num_threads=8, num_tasks=8","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5950"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.113270","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for fetches with num_threads=16, num_tasks=16","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5960"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.114684","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Initialized background executor for common operations (e.g. clearing old parts) with num_threads=8, num_tasks=8","source_file":"src\/Interpreters\/Context.cpp; void DB::Context::initializeBackgroundExecutorsIfNeeded()","source_line":"5970"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119071","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Context","message":"Server logging level is set to 'test' and performance is degraded. This cannot be used in production.","source_file":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","source_line":"932"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119653","thread_name":"","thread_id":"1","level":"Warning","query_id":"","logger_name":"Context","message":"Delay accounting is not enabled, OSIOWaitMicroseconds will not be gathered. You can enable it using `echo 1 > \/proc\/sys\/kernel\/task_delayacct` or by using sysctl.","source_file":"src\/Interpreters\/Context.cpp; void DB::ContextSharedPart::addWarningMessage(const String &)","source_line":"932"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119893","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DNSCacheUpdater","message":"Update period 15 seconds","source_file":"src\/Interpreters\/DNSCacheUpdater.cpp; void DB::DNSCacheUpdater::start()","source_line":"49"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119913","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Loading metadata from \/var\/lib\/clickhouse\/","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2315"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119951","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Database disk name: default","source_file":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","source_line":"1182"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119958","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Context","message":"Database disk name: default, path: \/var\/lib\/clickhouse\/","source_file":"src\/Interpreters\/Context.cpp; auto DB::Context::getDatabaseDisk()::(anonymous class)::operator()() const","source_line":"1190"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.119969","thread_name":"","thread_id":"17","level":"Debug","query_id":"","logger_name":"DNSResolver","message":"Updating DNS cache","source_file":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","source_line":"446"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.120065","thread_name":"","thread_id":"17","level":"Debug","query_id":"","logger_name":"DNSResolver","message":"Updated DNS cache","source_file":"src\/Common\/DNSResolver.cpp; bool DB::DNSResolver::updateCache(UInt32)","source_line":"477"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.120009","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/system.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.123722","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/information_schema.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.125439","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"loadSystemDatabase","message":"metadata_file_path metadata\/INFORMATION_SCHEMA.sql, existsFile false","source_file":"src\/Interpreters\/loadMetadata.cpp; void DB::loadSystemDatabaseImpl(ContextMutablePtr, const String &, const String &)","source_line":"305"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126783","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Metadata processed, database system has 0 tables and 0 dictionaries in total.","source_file":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","source_line":"311"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126867","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"TablesLoader","message":"Parsed metadata of 0 tables in 1 databases in 6.4481e-05 sec","source_file":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","source_line":"63"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126892","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ReferentialDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126902","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"LoadingDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126949","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Prioritize load job 'startup Atomic database system': BackgrndStartup -> ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","source_line":"693"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.126995","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Prioritize load job 'startup Ordinary database system': BackgrndStartup -> ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::prioritize(const LoadJobPtr &, size_t, std::unique_lock &)","source_line":"693"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127019","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Ordinary database system' into ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127029","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Atomic database system' into ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127040","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: none -> 0","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127049","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #1 in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127173","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Wait load job 'startup Atomic database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::wait(std::unique_lock &, const LoadJobPtr &)","source_line":"802"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127250","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Ordinary database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127325","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Ordinary database system' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127350","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #2 in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127685","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Atomic database system' in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127724","thread_name":"","thread_id":"648","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127818","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Atomic database system' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127830","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in ForegroundLoad","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127837","thread_name":"","thread_id":"647","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: 0 -> none","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.127862","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_log from query_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.129919","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_thread_log from query_thread_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryThreadLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.130583","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.part_log from part_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::PartLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.131104","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.trace_log from trace_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TraceLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.131563","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.crash_log from crash_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::CrashLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.132503","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.text_log from text_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TextLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.133105","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.metric_log from metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::MetricLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.140638","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.latency_log from latency_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::LatencyLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.140928","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.error_log from error_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ErrorLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141147","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.filesystem_cache_log since corresponding section 'filesystem_cache_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemCacheLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141221","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.filesystem_read_prefetches_log since corresponding section 'filesystem_read_prefetches_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::FilesystemReadPrefetchesLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141242","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.s3queue_log from s3queue_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141497","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.azure_queue_log since corresponding section 'azure_queue_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ObjectStorageQueueLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141516","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.asynchronous_metric_log from asynchronous_metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousMetricLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141715","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.opentelemetry_span_log from opentelemetry_span_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::OpenTelemetrySpanLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.141990","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_views_log from query_views_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryViewsLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142345","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.zookeeper_log since corresponding section 'zookeeper_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ZooKeeperLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142372","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.session_log since corresponding section 'session_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::SessionLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142385","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Not creating system.transactions_info_log since corresponding section 'transactions_info_log' is missing from config","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::TransactionsInfoLog]","source_line":"153"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142397","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.processors_profile_log from processors_profile_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::ProcessorsProfileLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.142770","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.asynchronous_insert_log from asynchronous_insert_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::AsynchronousInsertLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143132","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.backup_log from backup_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BackupLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143443","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.blob_storage_log from blob_storage_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::BlobStorageLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.143675","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"SystemLog","message":"Creating system.query_metric_log from query_metric_log","source_file":"src\/Interpreters\/SystemLog.cpp; std::shared_ptr DB::(anonymous namespace)::createSystemLog(ContextPtr, const String &, const String &, const Poco::Util::AbstractConfiguration &, const String &, const String &) [TSystemLog = DB::QueryMetricLog]","source_line":"158"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.149124","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.149151","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150446","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150457","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150483","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150489","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150508","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150516","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150535","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150540","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150582","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150593","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150693","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150700","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150750","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150756","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150813","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150821","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150866","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150872","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150907","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150913","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150957","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150963","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.150997","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151003","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151041","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151047","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151088","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151095","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151137","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151143","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151168","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151178","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151202","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151207","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151230","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151272","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151323","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151330","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151353","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151358","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151390","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151400","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151480","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151488","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151524","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151569","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151604","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151609","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151681","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151687","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151743","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151750","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151919","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151927","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151959","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151965","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151989","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.151995","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152021","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152057","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152064","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152118","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152124","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152163","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152169","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152221","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152226","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152280","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152286","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152422","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152431","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152563","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152578","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152628","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152634","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152760","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152767","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152796","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152803","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152823","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152828","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152852","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152863","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152937","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152945","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152971","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152977","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.152996","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153000","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153048","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153054","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153092","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153099","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153134","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153141","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153331","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153339","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153437","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153444","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153481","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153486","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153505","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153512","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153555","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153564","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153611","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153617","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153640","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153645","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153665","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153669","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153851","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.153862","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154005","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154017","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154080","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154086","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154253","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154262","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154381","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154388","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154438","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154445","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154491","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154496","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154663","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154671","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154729","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154735","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154784","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154790","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154824","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154830","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154872","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.154878","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155016","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155027","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155085","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155091","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155149","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155156","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155193","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155199","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155264","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155270","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155319","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155329","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155387","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155394","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155458","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155466","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155491","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155497","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155528","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155533","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155573","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155583","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155619","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155625","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155660","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155694","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155778","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155785","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155826","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155832","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155859","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155864","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155907","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155912","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155941","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.155946","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156005","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156011","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156089","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156096","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156123","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156128","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156163","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156169","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156201","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156207","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156235","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156240","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156269","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156274","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156295","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156314","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156363","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156373","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156404","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156409","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156435","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156440","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156521","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.156527","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.163703","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseCatalog","message":"Found 0 partially dropped tables. Will load them and retry removal.","source_file":"src\/Interpreters\/DatabaseCatalog.cpp; void DB::DatabaseCatalog::loadMarkedAsDroppedTables()","source_line":"1050"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165325","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"DatabaseAtomic (default)","message":"Metadata processed, database default has 0 tables and 0 dictionaries in total.","source_file":"src\/Databases\/DatabaseOrdinary.cpp; virtual void DB::DatabaseOrdinary::loadTablesMetadata(ContextPtr, ParsedTablesMetadata &, bool)","source_line":"311"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165356","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"TablesLoader","message":"Parsed metadata of 0 tables in 1 databases in 9.2864e-05 sec","source_file":"src\/Databases\/TablesLoader.cpp; LoadTaskPtrs DB::TablesLoader::loadTablesAsync(LoadJobSet)","source_line":"63"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165381","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"ReferentialDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165389","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"LoadingDeps","message":"No tables","source_file":"src\/Databases\/TablesDependencyGraph.cpp; void DB::TablesDependencyGraph::log() const","source_line":"722"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165432","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"loadMetadata","message":"Start asynchronous loading of databases","source_file":"src\/Interpreters\/loadMetadata.cpp; LoadTaskPtrs DB::loadMetadata(ContextMutablePtr, const String &, bool)","source_line":"268"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165454","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Ordinary database default' into BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165463","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Schedule load job 'startup Atomic database default' into BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::schedule(const LoadJobSet &)","source_line":"347"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165474","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: none -> 2","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165482","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #1 in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165670","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"UserDefinedSQLObjectsLoaderFromDisk","message":"Loading user defined objects from \/var\/lib\/clickhouse\/user_defined\/","source_file":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","source_line":"131"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165691","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"UserDefinedSQLObjectsLoaderFromDisk","message":"The directory for user defined objects (\/var\/lib\/clickhouse\/user_defined\/) does not exist: nothing to load","source_file":"src\/Functions\/UserDefined\/UserDefinedSQLObjectsDiskStorage.cpp; void DB::UserDefinedSQLObjectsDiskStorage::loadObjectsImpl()","source_line":"135"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165690","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Ordinary database default' in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165741","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Ordinary database default' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165748","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"WorkloadEntityDiskStorage","message":"Loading workload entities from \/var\/lib\/clickhouse\/workload\/","source_file":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","source_line":"129"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165762","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Spawn loader worker #2 in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::spawn(Pool &, std::unique_lock &)","source_line":"883"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165787","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"WorkloadEntityDiskStorage","message":"The directory for workload entities (\/var\/lib\/clickhouse\/workload\/) does not exist: nothing to load","source_file":"src\/Common\/Scheduler\/Workload\/WorkloadEntityDiskStorage.cpp; void DB::WorkloadEntityDiskStorage::loadEntitiesImpl()","source_line":"133"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165801","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"Application","message":"Loaded metadata.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2408"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165876","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Execute load job 'startup Atomic database default' in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"934"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165887","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Finish load job 'startup Atomic database default' with status OK","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::finish(const LoadJobPtr &, LoadStatus, std::exception_ptr, std::unique_lock &)","source_line":"606"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165893","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165903","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Stop worker in BackgrndStartup","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::worker(Pool &)","source_line":"916"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165915","thread_name":"","thread_id":"668","level":"Debug","query_id":"","logger_name":"AsyncLoader","message":"Change current priority: 2 -> none","source_file":"src\/Common\/AsyncLoader.cpp; void DB::AsyncLoader::setCurrentPriority(std::unique_lock &, std::optional)","source_line":"850"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165957","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"bool DB::(anonymous namespace)::checkPermissionsImpl()","message":"Code: 412. DB::Exception: Can't receive Netlink response: error -2. (NETLINK_ERROR) (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Common\/NetlinkMetricsProvider.cpp; bool DB::(anonymous namespace)::checkPermissionsImpl()","source_line":"216"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.165987","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Tasks stats provider: procfs","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2428"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166076","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"There are 0 detached tables. Start searching non used tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"440"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166089","thread_name":"","thread_id":"1","level":"Debug","query_id":"","logger_name":"DatabaseAtomic (system)","message":"Found 0 non used tables in detached tables.","source_file":"src\/Databases\/DatabaseAtomic.cpp; DatabaseAtomic::DetachedTables DB::DatabaseAtomic::cleanupDetachedTables()","source_line":"451"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166466","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Failed to create SSL context. SSL will be disabled. Error: Poco::Exception. Code: 1000, e.code() = 0, SSL Exception: Configuration error: no certificate file has been specified (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","source_line":"37"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166561","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Failed to read RSA key pair from server certificate. Error: Code: 139. DB::Exception: Certificate file is not set. (NO_ELEMENTS_IN_CONFIG) (version 25.2.2.96963.altinityantalya.96963 (official build))","source_file":"src\/Server\/MySQLHandlerFactory.cpp; DB::MySQLHandlerFactory::MySQLHandlerFactory(IServer &, const ProfileEvents::Event &, const ProfileEvents::Event &)","source_line":"48"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.166575","thread_name":"","thread_id":"1","level":"Trace","query_id":"","logger_name":"MySQLHandlerFactory","message":"Generating new RSA key pair.","source_file":"src\/Server\/MySQLHandlerFactory.cpp; void DB::MySQLHandlerFactory::generateRSAKeys()","source_line":"107"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.222054","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CertificateReloader","message":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","source_file":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","source_line":"142"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.222082","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"CertificateReloader","message":"One of paths is empty. Cannot apply new configuration for certificates. Fill all paths and try again.","source_file":"src\/Server\/CertificateReloader.cpp; void DB::CertificateReloader::tryLoadImpl(const Poco::Util::AbstractConfiguration &, SSL_CTX *, const std::string &)","source_line":"142"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225288","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for http:\/\/0.0.0.0:8123","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225355","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for native protocol (tcp): 0.0.0.0:9000","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225405","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for MySQL compatibility protocol: 0.0.0.0:9004","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225460","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Listening for PostgreSQL compatibility protocol: 0.0.0.0:9005","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2544"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.225467","thread_name":"","thread_id":"1","level":"Information","query_id":"","logger_name":"Application","message":"Ready for connections.","source_file":"programs\/server\/Server.cpp; virtual int DB::Server::main(const std::vector &)","source_line":"2548"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246774","thread_name":"TCPServer: 0.0.0.0:9000","thread_id":"675","level":"Test","query_id":"","logger_name":"Poco","message":"Queue size: 0, current threads: 0, threads in pool: 3, current connections: 0","source_file":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","source_line":"49"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246896","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCPHandlerFactory","message":"TCP Request. Address: 172.16.1.1:64780","source_file":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","source_line":"54"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246975","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Information","query_id":"","logger_name":"TCPHandler","message":"Client has not sent any data.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","source_line":"330"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.246986","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Done processing connection.","source_file":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","source_line":"2630"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774050","thread_name":"TCPServer: 0.0.0.0:9000","thread_id":"675","level":"Test","query_id":"","logger_name":"Poco","message":"Queue size: 0, current threads: 1, threads in pool: 3, current connections: 0","source_file":"src\/Common\/ErrorHandlers.h; virtual void ServerErrorHandler::logMessageImpl(Poco::Message::Priority, const std::string &)","source_line":"49"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774109","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCPHandlerFactory","message":"TCP Request. Address: 172.16.1.1:64794","source_file":"src\/Server\/TCPHandlerFactory.h; virtual Poco::Net::TCPServerConnection *DB::TCPHandlerFactory::createConnection(const Poco::Net::StreamSocket &, TCPServer &)","source_line":"54"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774175","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Connected ClickHouse client version 25.2.0, revision: 54476, user: default.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::receiveHello()","source_line":"1695"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774206","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"Authenticating user 'default' from 172.16.1.1:64794","source_file":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","source_line":"364"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774269","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Authenticated with global context as user 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; void DB::Session::authenticate(const Credentials &, const Poco::Net::SocketAddress &, const Strings &)","source_line":"374"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774291","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"Creating session context with user_id: 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeSessionContext()","source_line":"534"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774407","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"Settings: readonly = 0, allow_ddl = true, allow_introspection_functions = false","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"432"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774440","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"List of all grants: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"433"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.774461","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"ContextAccess (default)","message":"List of all grants including implicit: GRANT CHECK, SHOW, SELECT, INSERT, ALTER, CREATE, DROP, UNDROP TABLE, TRUNCATE, OPTIMIZE, BACKUP, KILL QUERY, KILL TRANSACTION, MOVE PARTITION BETWEEN SHARDS, ROLE ADMIN, CREATE ROW POLICY, ALTER ROW POLICY, DROP ROW POLICY, CREATE QUOTA, ALTER QUOTA, DROP QUOTA, CREATE SETTINGS PROFILE, ALTER SETTINGS PROFILE, DROP SETTINGS PROFILE, ALLOW SQL SECURITY NONE, SHOW ACCESS, SYSTEM, dictGet, displaySecretsInShowAndSelect, INTROSPECTION, SOURCES, CLUSTER ON *.* WITH GRANT OPTION, GRANT TABLE ENGINE ON * WITH GRANT OPTION, GRANT CREATE USER, ALTER USER, DROP USER, CREATE ROLE, ALTER ROLE, DROP ROLE, SET DEFINER ON * WITH GRANT OPTION, GRANT ALTER NAMED COLLECTION, CREATE NAMED COLLECTION, DROP NAMED COLLECTION, SHOW NAMED COLLECTIONS, NAMED COLLECTION ON * WITH GRANT OPTION","source_file":"src\/Access\/ContextAccess.cpp; void DB::ContextAccess::calculateAccessRights() const","source_line":"434"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784574","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Creating query context from session context, user_id: 94309d50-4f52-5250-31bd-74fecac179db, parent context user: default","source_file":"src\/Interpreters\/Session.cpp; ContextMutablePtr DB::Session::makeQueryContextImpl(const ClientInfo *, ClientInfo *) const","source_line":"664"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784865","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"executeQuery","message":"(from 172.16.1.1:64794) (query 1, line 1) SELECT 1 (stage: Complete)","source_file":"src\/Interpreters\/executeQuery.cpp; void DB::logQuery(const String &, ContextPtr, bool, QueryProcessingStage::Enum)","source_line":"237"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.784896","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"CancellationChecker","message":"Did not add the task because the timeout is 0. Query: SELECT 1","source_file":"src\/Interpreters\/CancellationChecker.cpp; void DB::CancellationChecker::appendTask(const std::shared_ptr &, const Int64 &, OverflowMode)","source_line":"77"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785063","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"Planner","message":"Query to stage Complete","source_file":"src\/Planner\/Planner.cpp; void DB::Planner::buildQueryPlanIfNeeded()","source_line":"1296"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785140","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Trace","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"Planner","message":"Query from stage FetchColumns to stage Complete","source_file":"src\/Planner\/Planner.cpp; void DB::Planner::buildPlanForQueryNode()","source_line":"1574"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785260","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"QueryMetricLog","message":"Scheduling next collecting task for query_id 1059a1fa-69ac-4a60-8029-f5334c480773 in 999 ms","source_file":"src\/Interpreters\/QueryMetricLog.cpp; void DB::QueryMetricLogStatus::scheduleNext(String)","source_line":"219"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.785920","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"executeQuery","message":"Read 1 rows, 1.00 B in 0.001067 sec., 937.207122774133 rows\/sec., 937.21 B\/sec.","source_file":"src\/Interpreters\/executeQuery.cpp; void DB::logQueryFinish(QueryLogElement &, const ContextMutablePtr &, const ASTPtr &, const QueryPipeline &, bool, std::shared_ptr, QueryCacheUsage, bool)","source_line":"592"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.786033","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"1059a1fa-69ac-4a60-8029-f5334c480773","logger_name":"TCPHandler","message":"Processed in 0.001574242 sec.","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::logQueryDuration(QueryState &)","source_line":"865"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795736","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Test","query_id":"","logger_name":"TCPHandler","message":"Closing connection (open: true, cancelled: false, eof: true)","source_file":"src\/Server\/TCPHandler.cpp; void DB::TCPHandler::runImpl()","source_line":"456"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795753","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCPHandler","message":"Done processing connection.","source_file":"src\/Server\/TCPHandler.cpp; virtual void DB::TCPHandler::run()","source_line":"2630"} {"date_time_utc":"2025-04-02T04:15:43Z","date_time":"1743567343.795778","thread_name":"TCPServerConnection ([#1])","thread_id":"11","level":"Debug","query_id":"","logger_name":"TCP-Session-f4b7d46e-20aa-4d32-8683-e18e6bd888ca","message":"f4b7d46e-20aa-4d32-8683-e18e6bd888ca Logout, user_id: 94309d50-4f52-5250-31bd-74fecac179db","source_file":"src\/Interpreters\/Session.cpp; DB::Session::~Session()","source_line":"308"} run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/store/f64/f642cc61-bbfa-4244-909e-9f2838edf6cc/', '-type', 'f', '-name', '*.bin', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/store/f64/f642cc61-bbfa-4244-909e-9f2838edf6cc/', '-type', 'f', '-name', '*.bin', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/store/f64/f642cc61-bbfa-4244-909e-9f2838edf6cc/ -type f -name *.bin -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/store/f64/f642cc61-bbfa-4244-909e-9f2838edf6cc/ -type f -name *.bin -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottests3tablefunctions-gw5 No running containers for project: roottests3tablefunctions-gw5 Trying to prune unused networks... Trying to prune unused networks... [gw8] PASSED test_structured_logging_json/test.py::test_structured_logging_json_format Command:[docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml stop --timeout 20] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (89, '0x1ef1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (89, '0x1ef1') on node Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:9 Stdout:9 Command:[docker volume prune -f] Command:[docker volume prune -f] Executing query ALTER TABLE zero_copy_mutation ADD COLUMN valueX String MATERIALIZED value1 on node1 Executing query ALTER TABLE zero_copy_mutation ADD COLUMN valueX String MATERIALIZED value1 on node1 Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 9 Volumes pruned: 9 Cluster stopped Cluster stopped test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data Running tests in /ClickHouse/tests/integration/test_temporary_data_in_cache/test.py Running tests in /ClickHouse/tests/integration/test_temporary_data_in_cache/test.py Cluster start called. is_up=False Cluster start called. is_up=False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Docker networks for project roottesttemporarydataincache-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttemporarydataincache-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttemporarydataincache-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttemporarydataincache-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (90, '0x1fa4') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (90, '0x1fa4') on node Executing query ALTER TABLE zero_copy_mutation MATERIALIZE COLUMN valueX on node1 Executing query ALTER TABLE zero_copy_mutation MATERIALIZE COLUMN valueX on node1 Docker volumes for project roottesttemporarydataincache-gw5 are DRIVER VOLUME NAME Docker volumes for project roottesttemporarydataincache-gw5 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottesttemporarydataincache-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttemporarydataincache-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttemporarydataincache-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttemporarydataincache-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttemporarydataincache-gw5 are DRIVER VOLUME NAME Docker volumes for project roottesttemporarydataincache-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttemporarydataincache-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttemporarydataincache-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttemporarydataincache-gw5 No running containers for project: roottesttemporarydataincache-gw5 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (91, '0x2059') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (91, '0x2059') on node Stdout:9 Stdout:9 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 9 Volumes pruned: 9 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_temporary_data_in_cache/configs/config.d/storage_configuration.xml'] to /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_temporary_data_in_cache/configs/config.d/storage_configuration.xml'] to /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/database Setup database dir /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/database Setup logs dir /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/logs Setup logs dir /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml pull] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (92, '0x2110') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (92, '0x2110') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (93, '0x21c9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (93, '0x21c9') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (94, '0x2284') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (94, '0x2284') on node Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (95, '0x2341') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (95, '0x2341') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (96, '0x2400') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (96, '0x2400') on node run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (97, '0x24c1') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (97, '0x24c1') on node Stdout:9 Stdout:9 Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (98, '0x2584') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (98, '0x2584') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (99, '0x2649') on node Executing query INSERT INTO FUNCTION mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') (key, data) VALUES (99, '0x2649') on node Executing query SELECT COUNT() FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT COUNT() FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT sum(key) FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT sum(key) FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') on node Executing query SELECT sum(key) FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', 'key UInt64, data String') on node Executing query SELECT sum(key) FROM mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', 'key UInt64, data String') on node Executing query SELECT data from mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') where key = 42 on node Executing query SELECT data from mongodb('mongo1:27017', 'test', 'simple_table', 'root', 'clickhouse', structure='key UInt64, data String') where key = 42 on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml stop --timeout 20] [gw3] PASSED test_table_function_mongodb_legacy/test.py::test_simple_select[False] Executing query SELECT count() FROM system.mutations WHERE table='zero_copy_mutation' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='zero_copy_mutation' AND is_done=0 on node1 Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestsystemlogscomment-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/containers/roottestsystemlogscomment-gw9-node_default-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/406d7b1c14cd54eb1fcb9f69d2c7c3dae3d377cf26425b4446a02fd60148a6cb/start HTTP/1.1" 200 0 http://localhost:None "POST /v1.46/exec/406d7b1c14cd54eb1fcb9f69d2c7c3dae3d377cf26425b4446a02fd60148a6cb/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/406d7b1c14cd54eb1fcb9f69d2c7c3dae3d377cf26425b4446a02fd60148a6cb/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/exec/406d7b1c14cd54eb1fcb9f69d2c7c3dae3d377cf26425b4446a02fd60148a6cb/json HTTP/1.1" 200 586 Executing query SELECT count() FROM system.mutations WHERE table='zero_copy_mutation' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='zero_copy_mutation' AND is_done=0 on node2 Executing query SELECT arrayJoin(data_paths) FROM system.tables WHERE name='zero_copy_mutation' on node2 Executing query SELECT arrayJoin(data_paths) FROM system.tables WHERE name='zero_copy_mutation' on node2 run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/store/bd9/bd9df950-da7d-4574-ac43-0f912ddb60cd/', '-type', 'f', '-name', '*.bin', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/store/bd9/bd9df950-da7d-4574-ac43-0f912ddb60cd/', '-type', 'f', '-name', '*.bin', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/store/bd9/bd9df950-da7d-4574-ac43-0f912ddb60cd/ -type f -name *.bin -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/store/bd9/bd9df950-da7d-4574-ac43-0f912ddb60cd/ -type f -name *.bin -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Executing query ALTER TABLE zero_copy_mutation DETACH PARTITION '0' on node1 Executing query ALTER TABLE zero_copy_mutation DETACH PARTITION '0' on node1 Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Executing query ALTER TABLE zero_copy_mutation ATTACH PARTITION '0' on node1 Executing query ALTER TABLE zero_copy_mutation ATTACH PARTITION '0' on node1 Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.3.3, port:2181, use_ssl:False get_kazoo_client: zoo2, ip:172.16.3.3, port:2181, use_ssl:False Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-zoo3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.3.2, port:2181, use_ssl:False get_kazoo_client: zoo3, ip:172.16.3.2, port:2181, use_ssl:False Stdout:748 Stdout:748 Clickhouse process running. Clickhouse process running. run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] run container_id:roottestsystemlogscomment-gw9-node_default-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker exec roottestsystemlogscomment-gw9-node_default-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Stdout:748 Stdout:748 Executing query select 20 on node_default Executing query select 20 on node_default Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml up -d --no-recreate] Executing query select 1 on node_default Executing query select 1 on node_default Executing query system flush logs on node_default Executing query system flush logs on node_default Executing query SELECT comment FROM system.tables WHERE name = 'query_log' on node_default Executing query SELECT comment FROM system.tables WHERE name = 'query_log' on node_default Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml stop --timeout 20] [gw9] PASSED test_system_logs_comment/test.py::test_system_logs_comment Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Running Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Running Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Running Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Running Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Running Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Running Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Creating Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Created Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Starting Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Started Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.3.7... Waiting for ClickHouse start in node1, ip: 172.16.3.7... http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Stopping Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Stopping Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Stopped Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/.env --project-name roottestsystemlogscomment-gw9 --file /ClickHouse/tests/integration/test_system_logs_comment/_instances-0-gw9/node_default/docker-compose.yml down --volumes] http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None Executing query SELECT count() FROM system.mutations WHERE table='zero_copy_mutation' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='zero_copy_mutation' AND is_done=0 on node1 http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/.env --project-name rootteststructuredloggingjson-gw8 --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_all_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_some_keys/docker-compose.yml --file /ClickHouse/tests/integration/test_structured_logging_json/_instances-0-gw8/node_no_keys/docker-compose.yml down --volumes] http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dd899b256adb0c88fbde512e7b7b4508151903b09cf9c564a613c25c1b7b61b9/json HTTP/1.1" 200 None ClickHouse node1 started ClickHouse node1 started get_instance_ip instance_name=node2 get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.3.5... Waiting for ClickHouse start in node2, ip: 172.16.3.5... http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/520dccf9ffdf0db980c25c51fb7456f25e5e6a3a120a6eb5e74851ea77f093d0/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/520dccf9ffdf0db980c25c51fb7456f25e5e6a3a120a6eb5e74851ea77f093d0/json HTTP/1.1" 200 None ClickHouse node2 started ClickHouse node2 started get_instance_ip instance_name=node3 get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.3.6... Waiting for ClickHouse start in node3, ip: 172.16.3.6... http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemlogshostnamereplicated-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/db56d052561507de62314198f6d7f107147bbdd6b796f3e4898105e280a7ee15/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/db56d052561507de62314198f6d7f107147bbdd6b796f3e4898105e280a7ee15/json HTTP/1.1" 200 None ClickHouse node3 started ClickHouse node3 started Executing query CREATE DATABASE test; CREATE TABLE test.test_table(date Date, id UInt32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/1/replicated/test_table', 'node1') ORDER BY id PARTITION BY toYYYYMM(date); on node1 Executing query CREATE DATABASE test; CREATE TABLE test.test_table(date Date, id UInt32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/1/replicated/test_table', 'node1') ORDER BY id PARTITION BY toYYYYMM(date); on node1 Executing query CREATE DATABASE test; CREATE TABLE test.test_table(date Date, id UInt32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/1/replicated/test_table', 'node2') ORDER BY id PARTITION BY toYYYYMM(date); on node2 Executing query CREATE DATABASE test; CREATE TABLE test.test_table(date Date, id UInt32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/1/replicated/test_table', 'node2') ORDER BY id PARTITION BY toYYYYMM(date); on node2 Executing query CREATE DATABASE test; CREATE TABLE test.test_table(date Date, id UInt32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/1/replicated/test_table', 'node3') ORDER BY id PARTITION BY toYYYYMM(date); on node3 Executing query CREATE DATABASE test; CREATE TABLE test.test_table(date Date, id UInt32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/test/1/replicated/test_table', 'node3') ORDER BY id PARTITION BY toYYYYMM(date); on node3 Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Stopping Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Stopping Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Stopped Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Stopped Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Removing Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Removing Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Removed Stderr: Container roottestsystemlogscomment-gw9-node_default-1 Removed Stderr: Network roottestsystemlogscomment-gw9_default Removing Stderr: Network roottestsystemlogscomment-gw9_default Removing Stderr: Network roottestsystemlogscomment-gw9_default Removed Stderr: Network roottestsystemlogscomment-gw9_default Removed Cleanup called Cleanup called Executing query SELECT 1 on node1 Executing query SELECT 1 on node1 Docker networks for project roottestsystemlogscomment-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogscomment-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemlogscomment-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogscomment-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemlogscomment-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogscomment-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemlogscomment-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemlogscomment-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemlogscomment-gw9 No running containers for project: roottestsystemlogscomment-gw9 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:9 Stdout:9 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 9 Volumes pruned: 9 Executing query SELECT 1 on node2 Executing query SELECT 1 on node2 Executing query SELECT 1 on node3 Executing query SELECT 1 on node3 Executing query SYSTEM FLUSH LOGS on node1 Executing query SYSTEM FLUSH LOGS on node1 Executing query SYSTEM FLUSH LOGS on node2 Executing query SYSTEM FLUSH LOGS on node2 Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Stopping Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Removing Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Removing Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Removing Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Removing Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Stopped Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Removing Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Removing Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Removed Stderr: Container rootteststructuredloggingjson-gw8-node_all_keys-1 Removed Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Removed Stderr: Container rootteststructuredloggingjson-gw8-node_no_keys-1 Removed Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Removed Stderr: Container rootteststructuredloggingjson-gw8-node_some_keys-1 Removed Stderr: Network rootteststructuredloggingjson-gw8_default Removing Stderr: Network rootteststructuredloggingjson-gw8_default Removing Stderr: Network rootteststructuredloggingjson-gw8_default Removed Stderr: Network rootteststructuredloggingjson-gw8_default Removed Cleanup called Cleanup called Executing query SELECT count() FROM system.mutations WHERE table='zero_copy_mutation' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='zero_copy_mutation' AND is_done=0 on node2 Docker networks for project rootteststructuredloggingjson-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststructuredloggingjson-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project rootteststructuredloggingjson-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststructuredloggingjson-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststructuredloggingjson-gw8 are DRIVER VOLUME NAME Docker volumes for project rootteststructuredloggingjson-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststructuredloggingjson-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststructuredloggingjson-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststructuredloggingjson-gw8 No running containers for project: rootteststructuredloggingjson-gw8 Trying to prune unused networks... Trying to prune unused networks... Executing query DROP TABLE zero_copy_mutation SYNC on node1 Executing query DROP TABLE zero_copy_mutation SYNC on node1 Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/.env --project-name roottesttablefunctionmongodblegacy-gw3 --file /ClickHouse/tests/integration/test_table_function_mongodb_legacy/_instances-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mongo.yml down --volumes] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:9 Stdout:9 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 9 Volumes pruned: 9 Executing query SYSTEM FLUSH LOGS on node3 Executing query SYSTEM FLUSH LOGS on node3 Executing query SELECT hostname from system.query_log where query_id='query_node1' LIMIT 1 on node1 Executing query SELECT hostname from system.query_log where query_id='query_node1' LIMIT 1 on node1 Executing query SELECT hostname from system.query_log where query_id='query_node2' LIMIT 1 on node2 Executing query SELECT hostname from system.query_log where query_id='query_node2' LIMIT 1 on node2 Executing query SELECT hostname from system.query_log where query_id='query_node3' LIMIT 1 on node3 Executing query SELECT hostname from system.query_log where query_id='query_node3' LIMIT 1 on node3 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml stop --timeout 20] Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopping [gw7] PASSED test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopping Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Stopped Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removing Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-node-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_no_cred-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo1-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removed Stderr: Container roottesttablefunctionmongodblegacy-gw3-mongo_secure-1 Removed Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removing Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removing Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removed Stderr: Network roottesttablefunctionmongodblegacy-gw3_default Removed Cleanup called Cleanup called Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttablefunctionmongodblegacy-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttablefunctionmongodblegacy-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttablefunctionmongodblegacy-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttablefunctionmongodblegacy-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttablefunctionmongodblegacy-gw3 No running containers for project: roottesttablefunctionmongodblegacy-gw3 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:3 Stdout:3 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 3 Volumes pruned: 3 test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance Running tests in /ClickHouse/tests/integration/test_tcp_handler_http_responses/test_case.py Running tests in /ClickHouse/tests/integration/test_tcp_handler_http_responses/test_case.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottesttcphandlerhttpresponsescase-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttcphandlerhttpresponsescase-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttcphandlerhttpresponsescase-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttcphandlerhttpresponsescase-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttcphandlerhttpresponsescase-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttcphandlerhttpresponsescase-gw3 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottesttcphandlerhttpresponsescase-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttcphandlerhttpresponsescase-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttcphandlerhttpresponsescase-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttcphandlerhttpresponsescase-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttcphandlerhttpresponsescase-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttcphandlerhttpresponsescase-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttcphandlerhttpresponsescase-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttcphandlerhttpresponsescase-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttcphandlerhttpresponsescase-gw3 No running containers for project: roottesttcphandlerhttpresponsescase-gw3 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:3 Stdout:3 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 3 Volumes pruned: 3 Setup directory for instance: node_with_http Setup directory for instance: node_with_http Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_tcp_handler_http_responses/configs/config.d/http-port-31337.xml'] to /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_tcp_handler_http_responses/configs/config.d/http-port-31337.xml'] to /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/configs/config.d Setup database dir /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/database Setup database dir /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/database Setup logs dir /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/logs Setup logs dir /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node_without_http Setup directory for instance: node_without_http Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_tcp_handler_http_responses/configs/config.d/no-http-port.xml'] to /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_tcp_handler_http_responses/configs/config.d/no-http-port.xml'] to /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/configs/config.d Setup database dir /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/database Setup database dir /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/database Setup logs dir /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/logs Setup logs dir /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml pull] Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/.env --project-name roottestsystemlogshostnamereplicated-gw7 --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_system_logs_hostname/_instances-replicated-0-gw7/node3/docker-compose.yml down --volumes] Stderr: node_without_http Skipped - Image is already being pulled by node_with_http Stderr: node_without_http Skipped - Image is already being pulled by node_with_http Stderr: node_with_http Pulling Stderr: node_with_http Pulling Stderr: node_with_http Pulled Stderr: node_with_http Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml up -d --no-recreate] Stderr: node Pulling Stderr: node Pulling Stderr: node Pulled Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml up -d --no-recreate] Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-node1-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-node2-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-node3-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Stopping Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Stopped Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Removing Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo3-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo2-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Removed Stderr: Container roottestsystemlogshostnamereplicated-gw7-zoo1-1 Removed Stderr: Network roottestsystemlogshostnamereplicated-gw7_default Removing Stderr: Network roottestsystemlogshostnamereplicated-gw7_default Removing Stderr: Network roottestsystemlogshostnamereplicated-gw7_default Removed Stderr: Network roottestsystemlogshostnamereplicated-gw7_default Removed Cleanup called Cleanup called Docker networks for project roottestsystemlogshostnamereplicated-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemlogshostnamereplicated-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemlogshostnamereplicated-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemlogshostnamereplicated-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemlogshostnamereplicated-gw7 are DRIVER VOLUME NAME Docker volumes for project roottestsystemlogshostnamereplicated-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemlogshostnamereplicated-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemlogshostnamereplicated-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemlogshostnamereplicated-gw7 No running containers for project: roottestsystemlogshostnamereplicated-gw7 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:3 Stdout:3 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 3 Volumes pruned: 3 Stderr: Network roottesttcphandlerhttpresponsescase-gw3_default Creating Stderr: Network roottesttcphandlerhttpresponsescase-gw3_default Creating Stderr: Network roottesttcphandlerhttpresponsescase-gw3_default Created Stderr: Network roottesttcphandlerhttpresponsescase-gw3_default Created Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Creating Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Creating Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Creating Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Creating Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Created Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Created Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Created Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Created Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Starting Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Starting Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Starting Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Starting Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Started Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Started Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Started Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node_with_http get_instance_ip instance_name=node_with_http http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_with_http-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_with_http-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_with_http get_instance_ip instance_name=node_with_http http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_with_http-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_with_http-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_with_http, ip: 172.16.1.2... Waiting for ClickHouse start in node_with_http, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_with_http-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_with_http-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None Stderr: Network roottesttemporarydataincache-gw5_default Creating Stderr: Network roottesttemporarydataincache-gw5_default Creating Stderr: Network roottesttemporarydataincache-gw5_default Created Stderr: Network roottesttemporarydataincache-gw5_default Created Stderr: Container roottesttemporarydataincache-gw5-node-1 Creating Stderr: Container roottesttemporarydataincache-gw5-node-1 Creating Stderr: Container roottesttemporarydataincache-gw5-node-1 Created Stderr: Container roottesttemporarydataincache-gw5-node-1 Created Stderr: Container roottesttemporarydataincache-gw5-node-1 Starting Stderr: Container roottesttemporarydataincache-gw5-node-1 Starting Stderr: Container roottesttemporarydataincache-gw5-node-1 Started Stderr: Container roottesttemporarydataincache-gw5-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottesttemporarydataincache-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttemporarydataincache-gw5-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottesttemporarydataincache-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttemporarydataincache-gw5-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.2.2... Waiting for ClickHouse start in node, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottesttemporarydataincache-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttemporarydataincache-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f56d27089779807ab808895424936dec8e63c9b1fcf327e9f9326d49763aeb7f/json HTTP/1.1" 200 None ClickHouse node_with_http started ClickHouse node_with_http started get_instance_ip instance_name=node_without_http get_instance_ip instance_name=node_without_http http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_without_http-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_without_http-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_without_http get_instance_ip instance_name=node_without_http http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_without_http-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_without_http-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_without_http, ip: 172.16.1.3... Waiting for ClickHouse start in node_without_http, ip: 172.16.1.3... http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_without_http-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottesttcphandlerhttpresponsescase-gw3-node_without_http-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9631690e957ed0ca9c23ea9c5a98cfcdf2f401ab2831071975a38681e98f3e55/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9631690e957ed0ca9c23ea9c5a98cfcdf2f401ab2831071975a38681e98f3e55/json HTTP/1.1" 200 None ClickHouse node_without_http started ClickHouse node_without_http started Starting new HTTP connection (1): 172.16.1.2:9000 Starting new HTTP connection (1): 172.16.1.2:9000 http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://172.16.1.2:9000 "GET / HTTP/1.1" 400 None http://172.16.1.2:9000 "GET / HTTP/1.1" 400 None Encoding detection: ascii is most likely the one. Encoding detection: ascii is most likely the one. [gw3] PASSED test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance Starting new HTTP connection (1): 172.16.1.3:9000 Starting new HTTP connection (1): 172.16.1.3:9000 http://172.16.1.3:9000 "POST / HTTP/1.1" 400 None http://172.16.1.3:9000 "POST / HTTP/1.1" 400 None Encoding detection: ascii is most likely the one. Encoding detection: ascii is most likely the one. Encoding detection: ascii is most likely the one. Encoding detection: ascii is most likely the one. Command:[docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml stop --timeout 20] [gw3] PASSED test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6daeb95547aafd5fca2d0f302a3492eab30755f377b976647aa1ba372407da2d/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Executing query SYSTEM DROP FILESYSTEM CACHE on node Executing query SYSTEM DROP FILESYSTEM CACHE on node Executing query DROP TABLE IF EXISTS t1 SYNC on node Executing query DROP TABLE IF EXISTS t1 SYNC on node Executing query SELECT sum(size) FROM system.filesystem_cache on node Executing query SELECT sum(size) FROM system.filesystem_cache on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query CREATE TABLE t1 (x UInt64 CODEC(NONE)) ENGINE = MergeTree ORDER BY x SETTINGS storage_policy = 'tiny_local_cache' on node Executing query CREATE TABLE t1 (x UInt64 CODEC(NONE)) ENGINE = MergeTree ORDER BY x SETTINGS storage_policy = 'tiny_local_cache' on node Executing query INSERT INTO t1 SELECT number FROM numbers(1024 * 1024) on node Executing query INSERT INTO t1 SELECT number FROM numbers(1024 * 1024) on node Executing query OPTIMIZE TABLE t1 FINAL on node Executing query OPTIMIZE TABLE t1 FINAL on node Executing query SYSTEM STOP MERGES t1 on node Executing query SYSTEM STOP MERGES t1 on node Executing query SELECT sum(x) FROM t1 on node Executing query SELECT sum(x) FROM t1 on node Executing query SELECT sum(size) FROM system.filesystem_cache on node Executing query SELECT sum(size) FROM system.filesystem_cache on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT ignore(*) FROM numbers(10 * 1024 * 1024) ORDER BY sipHash64(number) on node Executing query SELECT ignore(*) FROM numbers(10 * 1024 * 1024) ORDER BY sipHash64(number) on node Executing query SELECT sum(size) FROM system.filesystem_cache on node Executing query SELECT sum(size) FROM system.filesystem_cache on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT avg(x) FROM t1 on node Executing query SELECT avg(x) FROM t1 on node Executing query SELECT sum(size) FROM system.filesystem_cache on node Executing query SELECT sum(size) FROM system.filesystem_cache on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT randomPrintableASCII(1024) FROM numbers(8 * 1024) FORMAT TSV on node via HTTP interface Executing query SELECT randomPrintableASCII(1024) FROM numbers(8 * 1024) FORMAT TSV on node via HTTP interface Starting new HTTP connection (1): 172.16.2.2:8123 Starting new HTTP connection (1): 172.16.2.2:8123 http://172.16.2.2:8123 "GET /?buffer_size=0&wait_end_of_query=1&query=SELECT+randomPrintableASCII%281024%29+FROM+numbers%288+%2A+1024%29+FORMAT+TSV HTTP/1.1" 200 None http://172.16.2.2:8123 "GET /?buffer_size=0&wait_end_of_query=1&query=SELECT+randomPrintableASCII%281024%29+FROM+numbers%288+%2A+1024%29+FORMAT+TSV HTTP/1.1" 200 None Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT free_space FROM system.disks WHERE name = 'tiny_local_cache_local_disk' on node Executing query SELECT randomPrintableASCII(1024) FROM numbers(32 * 1024) FORMAT TSV on node via HTTP interface Executing query SELECT randomPrintableASCII(1024) FROM numbers(32 * 1024) FORMAT TSV on node via HTTP interface Starting new HTTP connection (1): 172.16.2.2:8123 Starting new HTTP connection (1): 172.16.2.2:8123 http://172.16.2.2:8123 "GET /?buffer_size=0&wait_end_of_query=1&query=SELECT+randomPrintableASCII%281024%29+FROM+numbers%2832+%2A+1024%29+FORMAT+TSV HTTP/1.1" 500 None http://172.16.2.2:8123 "GET /?buffer_size=0&wait_end_of_query=1&query=SELECT+randomPrintableASCII%281024%29+FROM+numbers%2832+%2A+1024%29+FORMAT+TSV HTTP/1.1" 500 None Executing query DROP TABLE IF EXISTS t1 SYNC on node Executing query DROP TABLE IF EXISTS t1 SYNC on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml stop --timeout 20] [gw5] PASSED test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data Executing query SELECT * FROM system.merges WHERE table = 'mutation_simple' and progress < 1 on node2 Executing query SELECT * FROM system.merges WHERE table = 'mutation_simple' and progress < 1 on node2 Executing query SELECT * FROM system.merges WHERE table = 'mutation_simple' and progress < 1 on node2 Executing query SELECT * FROM system.merges WHERE table = 'mutation_simple' and progress < 1 on node2 Executing query DROP TABLE test.mutation_simple on node1 Executing query DROP TABLE test.mutation_simple on node1 Executing query DROP TABLE test.mutation_simple on node2 Executing query DROP TABLE test.mutation_simple on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml stop --timeout 20] [gw6] PASSED test_system_merges/test.py::test_mutation_simple[replicated] Executing query SELECT count() FROM zero_copy_mutation WHERE value3 LIKE '%ab%' on node2 Executing query SELECT count() FROM zero_copy_mutation WHERE value3 LIKE '%ab%' on node2 Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Stopping Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Stopping Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Stopping Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Stopping Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Stopped Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Stopped Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Stopped Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/.env --project-name roottesttcphandlerhttpresponsescase-gw3 --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_with_http/docker-compose.yml --file /ClickHouse/tests/integration/test_tcp_handler_http_responses/_instances-case-0-gw3/node_without_http/docker-compose.yml down --volumes] Executing query DROP TABLE zero_copy_mutation SYNC on node2 Executing query DROP TABLE zero_copy_mutation SYNC on node2 Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Stopping Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Stopping Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Stopping Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Stopping Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Stopped Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Stopped Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Removing Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Removing Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Stopped Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Stopped Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Removing Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Removing Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Removed Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_without_http-1 Removed Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Removed Stderr: Container roottesttcphandlerhttpresponsescase-gw3-node_with_http-1 Removed Stderr: Network roottesttcphandlerhttpresponsescase-gw3_default Removing Stderr: Network roottesttcphandlerhttpresponsescase-gw3_default Removing Stderr: Network roottesttcphandlerhttpresponsescase-gw3_default Removed Stderr: Network roottesttcphandlerhttpresponsescase-gw3_default Removed Cleanup called Cleanup called Docker networks for project roottesttcphandlerhttpresponsescase-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttcphandlerhttpresponsescase-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttcphandlerhttpresponsescase-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttcphandlerhttpresponsescase-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttcphandlerhttpresponsescase-gw3 are DRIVER VOLUME NAME Docker volumes for project roottesttcphandlerhttpresponsescase-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttcphandlerhttpresponsescase-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttcphandlerhttpresponsescase-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttcphandlerhttpresponsescase-gw3 No running containers for project: roottesttcphandlerhttpresponsescase-gw3 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:3 Stdout:3 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 3 Volumes pruned: 3 Stderr: Container roottesttemporarydataincache-gw5-node-1 Stopping Stderr: Container roottesttemporarydataincache-gw5-node-1 Stopping Stderr: Container roottesttemporarydataincache-gw5-node-1 Stopped Stderr: Container roottesttemporarydataincache-gw5-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/.env --project-name roottesttemporarydataincache-gw5 --file /ClickHouse/tests/integration/test_temporary_data_in_cache/_instances-0-gw5/node/docker-compose.yml down --volumes] Stderr: Container roottesttemporarydataincache-gw5-node-1 Stopping Stderr: Container roottesttemporarydataincache-gw5-node-1 Stopping Stderr: Container roottesttemporarydataincache-gw5-node-1 Stopped Stderr: Container roottesttemporarydataincache-gw5-node-1 Stopped Stderr: Container roottesttemporarydataincache-gw5-node-1 Removing Stderr: Container roottesttemporarydataincache-gw5-node-1 Removing Stderr: Container roottesttemporarydataincache-gw5-node-1 Removed Stderr: Container roottesttemporarydataincache-gw5-node-1 Removed Stderr: Network roottesttemporarydataincache-gw5_default Removing Stderr: Network roottesttemporarydataincache-gw5_default Removing Stderr: Network roottesttemporarydataincache-gw5_default Removed Stderr: Network roottesttemporarydataincache-gw5_default Removed Cleanup called Cleanup called Docker networks for project roottesttemporarydataincache-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottesttemporarydataincache-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottesttemporarydataincache-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottesttemporarydataincache-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottesttemporarydataincache-gw5 are DRIVER VOLUME NAME Docker volumes for project roottesttemporarydataincache-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottesttemporarydataincache-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottesttemporarydataincache-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottesttemporarydataincache-gw5 No running containers for project: roottesttemporarydataincache-gw5 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:3 Stdout:3 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 3 Volumes pruned: 3 Stderr: Container roottestsystemmerges-gw6-node1-1 Stopping Stderr: Container roottestsystemmerges-gw6-node1-1 Stopping Stderr: Container roottestsystemmerges-gw6-node2-1 Stopping Stderr: Container roottestsystemmerges-gw6-node2-1 Stopping Stderr: Container roottestsystemmerges-gw6-node2-1 Stopped Stderr: Container roottestsystemmerges-gw6-node2-1 Stopped Stderr: Container roottestsystemmerges-gw6-node1-1 Stopped Stderr: Container roottestsystemmerges-gw6-node1-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo3-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo3-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo2-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo2-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo1-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo1-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo3-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo3-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo1-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo1-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo2-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/.env --project-name roottestsystemmerges-gw6 --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_system_merges/_instances-0-gw6/node2/docker-compose.yml down --volumes] Stderr: Container roottestsystemmerges-gw6-node1-1 Stopping Stderr: Container roottestsystemmerges-gw6-node1-1 Stopping Stderr: Container roottestsystemmerges-gw6-node2-1 Stopping Stderr: Container roottestsystemmerges-gw6-node2-1 Stopping Stderr: Container roottestsystemmerges-gw6-node1-1 Stopped Stderr: Container roottestsystemmerges-gw6-node1-1 Stopped Stderr: Container roottestsystemmerges-gw6-node1-1 Removing Stderr: Container roottestsystemmerges-gw6-node1-1 Removing Stderr: Container roottestsystemmerges-gw6-node2-1 Stopped Stderr: Container roottestsystemmerges-gw6-node2-1 Stopped Stderr: Container roottestsystemmerges-gw6-node2-1 Removing Stderr: Container roottestsystemmerges-gw6-node2-1 Removing Stderr: Container roottestsystemmerges-gw6-node2-1 Removed Stderr: Container roottestsystemmerges-gw6-node2-1 Removed Stderr: Container roottestsystemmerges-gw6-node1-1 Removed Stderr: Container roottestsystemmerges-gw6-node1-1 Removed Stderr: Container roottestsystemmerges-gw6-zoo2-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo2-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo3-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo3-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo1-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo1-1 Stopping Stderr: Container roottestsystemmerges-gw6-zoo2-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo2-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo2-1 Removing Stderr: Container roottestsystemmerges-gw6-zoo2-1 Removing Stderr: Container roottestsystemmerges-gw6-zoo3-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo3-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo3-1 Removing Stderr: Container roottestsystemmerges-gw6-zoo3-1 Removing Stderr: Container roottestsystemmerges-gw6-zoo1-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo1-1 Stopped Stderr: Container roottestsystemmerges-gw6-zoo1-1 Removing Stderr: Container roottestsystemmerges-gw6-zoo1-1 Removing Stderr: Container roottestsystemmerges-gw6-zoo1-1 Removed Stderr: Container roottestsystemmerges-gw6-zoo1-1 Removed Stderr: Container roottestsystemmerges-gw6-zoo3-1 Removed Stderr: Container roottestsystemmerges-gw6-zoo3-1 Removed Stderr: Container roottestsystemmerges-gw6-zoo2-1 Removed Stderr: Container roottestsystemmerges-gw6-zoo2-1 Removed Stderr: Network roottestsystemmerges-gw6_default Removing Stderr: Network roottestsystemmerges-gw6_default Removing Stderr: Network roottestsystemmerges-gw6_default Removed Stderr: Network roottestsystemmerges-gw6_default Removed Cleanup called Cleanup called Docker networks for project roottestsystemmerges-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemmerges-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemmerges-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemmerges-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemmerges-gw6 are DRIVER VOLUME NAME Docker volumes for project roottestsystemmerges-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemmerges-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemmerges-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemmerges-gw6 No running containers for project: roottestsystemmerges-gw6 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:3 Stdout:3 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 3 Volumes pruned: 3 test_system_clusters_actual_information/test.py::test Running tests in /ClickHouse/tests/integration/test_system_clusters_actual_information/test.py Running tests in /ClickHouse/tests/integration/test_system_clusters_actual_information/test.py Cluster start called. is_up=False Cluster start called. is_up=False Docker networks for project roottestsystemclustersactualinformation-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemclustersactualinformation-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemclustersactualinformation-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemclustersactualinformation-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemclustersactualinformation-gw6 are DRIVER VOLUME NAME Docker volumes for project roottestsystemclustersactualinformation-gw6 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker networks for project roottestsystemclustersactualinformation-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemclustersactualinformation-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemclustersactualinformation-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemclustersactualinformation-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemclustersactualinformation-gw6 are DRIVER VOLUME NAME Docker volumes for project roottestsystemclustersactualinformation-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemclustersactualinformation-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemclustersactualinformation-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemclustersactualinformation-gw6 No running containers for project: roottestsystemclustersactualinformation-gw6 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:3 Stdout:3 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 3 Volumes pruned: 3 Setup directory for instance: node Setup directory for instance: node Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_system_clusters_actual_information/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_system_clusters_actual_information/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/database Setup database dir /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/database Setup logs dir /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/logs Setup logs dir /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml pull] [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node1 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node1 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node2 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node2 Executing query CREATE TABLE unfreeze_test ON CLUSTER test_cluster (d UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/unfreeze_test', '{replica}') ORDER BY d SETTINGS storage_policy='s3' on node1 Executing query CREATE TABLE unfreeze_test ON CLUSTER test_cluster (d UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/unfreeze_test', '{replica}') ORDER BY d SETTINGS storage_policy='s3' on node1 Executing query INSERT INTO unfreeze_test VALUES (0) on node1 Executing query INSERT INTO unfreeze_test VALUES (0) on node1 Executing query select count() from system.parts where table = 'unfreeze_test' and active on node2 Executing query select count() from system.parts where table = 'unfreeze_test' and active on node2 Executing query ALTER TABLE unfreeze_test FREEZE WITH NAME 'freeze_backup1' on node1 Executing query ALTER TABLE unfreeze_test FREEZE WITH NAME 'freeze_backup1' on node1 Executing query ALTER TABLE unfreeze_test FREEZE WITH NAME 'freeze_backup2' on node2 Executing query ALTER TABLE unfreeze_test FREEZE WITH NAME 'freeze_backup2' on node2 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node2 run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Executing query TRUNCATE TABLE unfreeze_test on node1 Executing query TRUNCATE TABLE unfreeze_test on node1 Executing query SYSTEM SYNC REPLICA unfreeze_test on node2 Executing query SYSTEM SYNC REPLICA unfreeze_test on node2 run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Executing query ALTER TABLE unfreeze_test UNFREEZE WITH NAME 'freeze_backup1' on node1 Executing query ALTER TABLE unfreeze_test UNFREEZE WITH NAME 'freeze_backup1' on node1 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node1 Executing query ALTER TABLE unfreeze_test UNFREEZE WITH NAME 'freeze_backup2' on node2 Executing query ALTER TABLE unfreeze_test UNFREEZE WITH NAME 'freeze_backup2' on node2 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node2 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node1 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node1 Stderr: zoo1 Skipped - Image is already being pulled by node Stderr: zoo1 Skipped - Image is already being pulled by node Stderr: zoo2 Skipped - Image is already being pulled by node Stderr: zoo2 Skipped - Image is already being pulled by node Stderr: zoo3 Skipped - Image is already being pulled by node Stderr: zoo3 Skipped - Image is already being pulled by node Stderr: node Pulling Stderr: node Pulling Stderr: node Pulled Stderr: node Pulled Setup ZooKeeper Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/coordination'] Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/keeper3/coordination'] Command:[docker compose --project-name roottestsystemclustersactualinformation-gw6 --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Command:[docker compose --project-name roottestsystemclustersactualinformation-gw6 --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node2 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node2 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node1 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node1 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node2 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node2 Executing query CREATE TABLE unfreeze_test ON CLUSTER test_cluster (d UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/unfreeze_test', '{replica}') ORDER BY d SETTINGS storage_policy='s3' on node1 Executing query CREATE TABLE unfreeze_test ON CLUSTER test_cluster (d UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/unfreeze_test', '{replica}') ORDER BY d SETTINGS storage_policy='s3' on node1 Executing query INSERT INTO unfreeze_test VALUES (0) on node1 Executing query INSERT INTO unfreeze_test VALUES (0) on node1 Executing query select count() from system.parts where table = 'unfreeze_test' and active on node2 Executing query select count() from system.parts where table = 'unfreeze_test' and active on node2 Stderr:time="2025-04-02T04:16:14Z" level=trace msg="Docker Desktop integration not enabled" Stderr:time="2025-04-02T04:16:14Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestsystemclustersactualinformation-gw6_default Creating Stderr: Network roottestsystemclustersactualinformation-gw6_default Creating Stderr: Network roottestsystemclustersactualinformation-gw6_default Created Stderr: Network roottestsystemclustersactualinformation-gw6_default Created Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Creating Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Creating Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Creating Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Creating Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Creating Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Creating Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Created Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Created Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Created Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Created Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Created Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Created Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Starting Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Starting Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Starting Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Starting Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Starting Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Starting Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Started Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Started Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Started Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Started Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Started Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Started Stderr:time="2025-04-02T04:16:15Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:16:15Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:16:15Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T04:16:15Z" level=debug msg="otel error" error="" Wait ZooKeeper to start Wait ZooKeeper to start get_instance_ip instance_name=zoo1 get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-zoo1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.1.3, port:2181, use_ssl:False get_kazoo_client: zoo1, ip:172.16.1.3, port:2181, use_ssl:False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query ALTER TABLE unfreeze_test FREEZE WITH NAME 'freeze_backup1' on node1 Executing query ALTER TABLE unfreeze_test FREEZE WITH NAME 'freeze_backup1' on node1 Executing query ALTER TABLE unfreeze_test FREEZE WITH NAME 'freeze_backup2' on node2 Executing query ALTER TABLE unfreeze_test FREEZE WITH NAME 'freeze_backup2' on node2 Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node1 Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node2 run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Executing query TRUNCATE TABLE unfreeze_test on node1 Executing query TRUNCATE TABLE unfreeze_test on node1 Executing query SYSTEM SYNC REPLICA unfreeze_test on node2 Executing query SYSTEM SYNC REPLICA unfreeze_test on node2 run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node1-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node1-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup1/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)"] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 bash -c echo $(if [ -e '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store' ]; then echo 'yes'; else echo 'no'; fi)] Stdout:yes Stdout:yes run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] run container_id:roottests3zerocopyreplication-gw0-node2-1 detach:False nothrow:False cmd: ['find', '/var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store', '-type', 'f', '-exec', 'grep', '-o', 'r[01]\\{64\\}-file-[[:lower:]]\\{32\\}', '{}', ';'] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Command:[docker exec roottests3zerocopyreplication-gw0-node2-1 find /var/lib/clickhouse/disks/s31/shadow/freeze_backup2/store -type f -exec grep -o r[01]\{64\}-file-[[:lower:]]\{32\} {} ;] Executing query SYSTEM UNFREEZE WITH NAME 'freeze_backup1' on node1 Executing query SYSTEM UNFREEZE WITH NAME 'freeze_backup1' on node1 Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node1 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node1 Executing query SYSTEM UNFREEZE WITH NAME 'freeze_backup2' on node2 Executing query SYSTEM UNFREEZE WITH NAME 'freeze_backup2' on node2 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node2 Executing query SELECT count() FROM system.mutations WHERE table='unfreeze_test' AND is_done=0 on node2 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node1 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node1 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node2 Executing query DROP TABLE IF EXISTS unfreeze_test SYNC on node2 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10] Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.1.2, port:2181, use_ssl:False get_kazoo_client: zoo2, ip:172.16.1.2, port:2181, use_ssl:False Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-zoo3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.1.4, port:2181, use_ssl:False get_kazoo_client: zoo3, ip:172.16.1.4, port:2181, use_ssl:False Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Sending request(xid=2): Close() Connection dropped: socket connection broken Connection dropped: socket connection broken Transition to CONNECTING Transition to CONNECTING Zookeeper connection lost Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate] Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Running Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Running Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Running Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Running Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Running Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Running Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Creating Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Creating Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Created Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Created Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Starting Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Starting Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Started Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Started ClickHouse instance created ClickHouse instance created get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.5... Waiting for ClickHouse start in node, ip: 172.16.1.5... http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestsystemclustersactualinformation-gw6-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fc41e9dc2a78a9cdf3495c19a1978a21865ac5e55580d8a5bb6783c22eef270e/json HTTP/1.1" 200 None ClickHouse node started ClickHouse node started Executing query CREATE TABLE distributed (id UInt32, date Date) ENGINE = Distributed('test_cluster', 'default', 'replicated') on node Executing query CREATE TABLE distributed (id UInt32, date Date) ENGINE = Distributed('test_cluster', 'default', 'replicated') on node Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SYSTEM RELOAD CONFIG on node Executing query SYSTEM RELOAD CONFIG on node Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query SELECT count() FROM distributed SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node Executing query SELECT count() FROM distributed SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Stdout:2025.04.02 04:16:26.858780 [ 1292 ] {} DatabaseCatalog: Removing unused directory store/100/10000000-1000-4000-8000-000000000003/ from disk default Stdout:2025.04.02 04:16:26.858780 [ 1292 ] {} DatabaseCatalog: Removing unused directory store/100/10000000-1000-4000-8000-000000000003/ from disk default 1 log line(s) matching "Removing unused directory" appeared in a 59.516 seconds 1 log line(s) matching "Removing unused directory" appeared in a 59.516 seconds run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'directories from store\''] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'directories from store\''] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'directories from store'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'directories from store'] Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Stdout:2025.04.02 04:15:25.815420 [ 825 ] {} DatabaseCatalog: Cleaned up 3 directories from store/ on disk default Stdout:2025.04.02 04:15:25.815420 [ 825 ] {} DatabaseCatalog: Cleaned up 3 directories from store/ on disk default 1 log line(s) matching "directories from store" appeared in a 0.054 seconds 1 log line(s) matching "directories from store" appeared in a 0.054 seconds run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'Nothing to clean up from store/\''] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 \'Nothing to clean up from store/\''] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'Nothing to clean up from store/'] Command:[docker exec rootteststorecleanup-gw1-node1-1 bash -c timeout 90 tail -Fn1000000 "/var/log/clickhouse-server/clickhouse-server.log" | grep -Em 1 'Nothing to clean up from store/'] Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Stdout:2025.04.02 04:16:28.861516 [ 1306 ] {} DatabaseCatalog: Nothing to clean up from store/ on disk default Stdout:2025.04.02 04:16:28.861516 [ 1306 ] {} DatabaseCatalog: Nothing to clean up from store/ on disk default 1 log line(s) matching "Nothing to clean up from store/" appeared in a 1.944 seconds 1 log line(s) matching "Nothing to clean up from store/" appeared in a 1.944 seconds run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store] Stdout:0e7 Stdout:0e7 Stdout:100 Stdout:100 Stdout:200 Stdout:200 Stdout:242 Stdout:242 Stdout:2b9 Stdout:2b9 Stdout:300 Stdout:300 Stdout:456 Stdout:456 Stdout:6ab Stdout:6ab Stdout:704 Stdout:704 Stdout:809 Stdout:809 Stdout:cab Stdout:cab Stdout:cad Stdout:cad Stdout:eca Stdout:eca run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store] Stdout:total 52 Stdout:total 52 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 0e7 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 0e7 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:16 100 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:16 100 Stdout:drwxr-x--- 4 root root 4096 Apr 2 04:15 200 Stdout:drwxr-x--- 4 root root 4096 Apr 2 04:15 200 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 242 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 242 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 2b9 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 2b9 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 300 Stdout:drwxr-x--- 5 root root 4096 Apr 2 04:15 300 Stdout:drwxr-xr-x 2 root root 4096 Apr 2 04:16 456 Stdout:drwxr-xr-x 2 root root 4096 Apr 2 04:16 456 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 6ab Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 6ab Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 704 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 704 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 809 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 809 Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 cab Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 cab Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 cad Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 cad Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 eca Stdout:drwxr-x--- 3 root root 4096 Apr 2 04:15 eca run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/100'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/100'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/100] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/100] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/100'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/100'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/100] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/100] Stdout:total 0 Stdout:total 0 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/200'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/200'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/200] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/200] Stdout:20000000-1000-4000-8000-000000000001 Stdout:20000000-1000-4000-8000-000000000001 Stdout:20000000-1000-4000-8000-000000000002 Stdout:20000000-1000-4000-8000-000000000002 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/200'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/200'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/200] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/200] Stdout:total 8 Stdout:total 8 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 20000000-1000-4000-8000-000000000002 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/300'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/300'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/300] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/300] Stdout:30000000-1000-4000-8000-000000000001 Stdout:30000000-1000-4000-8000-000000000001 Stdout:30000000-1000-4000-8000-000000000002 Stdout:30000000-1000-4000-8000-000000000002 Stdout:30000000-1000-4000-8000-000000000003 Stdout:30000000-1000-4000-8000-000000000003 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/300'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/300'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/300] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/300] Stdout:total 12 Stdout:total 12 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000001 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000002 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 Stdout:drwxr-x--- 2 root root 4096 Apr 2 04:15 30000000-1000-4000-8000-000000000003 run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/456'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '/var/lib/clickhouse//store/456'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/456] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls /var/lib/clickhouse//store/456] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/456'] run container_id:rootteststorecleanup-gw1-node1-1 detach:False nothrow:False cmd: ['ls', '-l', '/var/lib/clickhouse//store/456'] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/456] Command:[docker exec rootteststorecleanup-gw1-node1-1 ls -l /var/lib/clickhouse//store/456] Stdout:total 0 Stdout:total 0 Executing query ATTACH TABLE db3.log2 on node1 Executing query ATTACH TABLE db3.log2 on node1 Executing query ATTACH DATABASE db2 on node1 Executing query ATTACH DATABASE db2 on node1 Executing query ATTACH TABLE db3.log on node1 Executing query ATTACH TABLE db3.log on node1 Executing query SELECT * FROM db3.log on node1 Executing query SELECT * FROM db3.log on node1 Executing query SELECT * FROM db3.log2 on node1 Executing query SELECT * FROM db3.log2 on node1 Executing query SELECT * FROM db2.log on node1 Executing query SELECT * FROM db2.log on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml stop --timeout 20] [gw1] PASSED test_store_cleanup/test.py::test_store_cleanup Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Stderr: Container rootteststorecleanup-gw1-node1-1 Stopping Stderr: Container rootteststorecleanup-gw1-node1-1 Stopping Stderr: Container rootteststorecleanup-gw1-node1-1 Stopped Stderr: Container rootteststorecleanup-gw1-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/.env --project-name rootteststorecleanup-gw1 --file /ClickHouse/tests/integration/test_store_cleanup/_instances-0-gw1/node1/docker-compose.yml down --volumes] Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Stderr: Container rootteststorecleanup-gw1-node1-1 Stopping Stderr: Container rootteststorecleanup-gw1-node1-1 Stopping Stderr: Container rootteststorecleanup-gw1-node1-1 Stopped Stderr: Container rootteststorecleanup-gw1-node1-1 Stopped Stderr: Container rootteststorecleanup-gw1-node1-1 Removing Stderr: Container rootteststorecleanup-gw1-node1-1 Removing Stderr: Container rootteststorecleanup-gw1-node1-1 Removed Stderr: Container rootteststorecleanup-gw1-node1-1 Removed Stderr: Network rootteststorecleanup-gw1_default Removing Stderr: Network rootteststorecleanup-gw1_default Removing Stderr: Network rootteststorecleanup-gw1_default Removed Stderr: Network rootteststorecleanup-gw1_default Removed Cleanup called Cleanup called Docker networks for project rootteststorecleanup-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project rootteststorecleanup-gw1 are NETWORK ID NAME DRIVER SCOPE Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Docker containers for project rootteststorecleanup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project rootteststorecleanup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project rootteststorecleanup-gw1 are DRIVER VOLUME NAME Docker volumes for project rootteststorecleanup-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/rootteststorecleanup-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/rootteststorecleanup-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: rootteststorecleanup-gw1 No running containers for project: rootteststorecleanup-gw1 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:3 Stdout:3 Command:[docker volume prune -f] Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Volumes pruned: 3 Volumes pruned: 3 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_delete_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_delete_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10] test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3] Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54 on node1 Executing query INSERT INTO ttl_delete_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54 on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_delete_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0) on node1 Executing query INSERT INTO ttl_delete_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0) on node1 Executing query INSERT INTO ttl_delete_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0) on node1 Executing query INSERT INTO ttl_delete_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0) on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54 on node1 Executing query INSERT INTO ttl_delete_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54 on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_delete_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0) on node1 Executing query INSERT INTO ttl_delete_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0) on node1 Executing query INSERT INTO ttl_delete_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0) on node1 Executing query INSERT INTO ttl_delete_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0) on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_delete_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_delete_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_delete_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54 on node1 Executing query INSERT INTO ttl_delete_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54 on node1 Executing query INSERT INTO ttl_delete_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0) on node1 Executing query INSERT INTO ttl_delete_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0) on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_delete_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0) on node1 Executing query INSERT INTO ttl_delete_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0) on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_delete_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node1 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SYSTEM SYNC REPLICA ttl_delete_test on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_delete_test FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_delete_test SYNC on node2 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3] test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10] Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10] test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3] Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered' on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3] test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10] Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (10, now() - INTERVAL 3 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query INSERT INTO ttl_move_test VALUES (11, now() - INTERVAL 1 DAY) on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node1 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query SELECT d FROM ttl_move_test ORDER BY d FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10] test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3] Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query CREATE TABLE ttl_move_test ON CLUSTER test_cluster (d UInt64, d1 DateTime) ENGINE=ReplicatedMergeTree('/clickhouse/tables/ttl_move_test', '{replica}') ORDER BY d TTL d1 + INTERVAL 2 DAY TO VOLUME 'external' SETTINGS storage_policy='tiered_copy' on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (0,1742947200.0),(1,1742947200.0),(2,1742947200.0),(3,1742947200.0),(4,1742947200.0),(5,1742947200.0),(6,1742947200.0),(7,1742947200.0),(8,1742947200.0),(9,1742947200.0),(10,1742947200.0),(11,1742947200.0),(12,1742947200.0),(13,1742947200.0),(14,1742947200.0),(15,1742947200.0),(16,1742947200.0),(17,1742947200.0),(18,1742947200.0),(19,1742947200.0),(20,1742947200.0),(21,1742947200.0),(22,1742947200.0),(23,1742947200.0),(24,1742947200.0),(25,1742947200.0),(26,1742947200.0),(27,1742947200.0),(28,1742947200.0),(29,1742947200.0),(30,1742947200.0),(31,1742947200.0),(32,1742947200.0),(33,1742947200.0),(34,1742947200.0),(35,1742947200.0),(36,1742947200.0),(37,1742947200.0),(38,1742947200.0),(39,1742947200.0),(40,1742947200.0),(41,1742947200.0),(42,1742947200.0),(43,1742947200.0),(44,1742947200.0),(45,1742947200.0),(46,1742947200.0),(47,1742947200.0),(48,1742947200.0),(49,1742947200.0),(50,1742947200.0),(51,1742947200.0),(52,1742947200.0),(53,1742947200.0),(54,1 on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (1048576,1743292800.0),(1048577,1743292800.0),(1048578,1743292800.0),(1048579,1743292800.0),(1048580,1743292800.0),(1048581,1743292800.0),(1048582,1743292800.0),(1048583,1743292800.0),(1048584,1743292800.0),(1048585,1743292800.0),(1048586,1743292800.0),(1048587,1743292800.0),(1048588,1743292800.0),(1048589,1743292800.0),(1048590,1743292800.0),(1048591,1743292800.0),(1048592,1743292800.0),(1048593,1743292800.0),(1048594,1743292800.0),(1048595,1743292800.0),(1048596,1743292800.0),(1048597,1743292800.0),(1048598,1743292800.0),(1048599,1743292800.0),(1048600,1743292800.0),(1048601,1743292800.0),(1048602,1743292800.0),(1048603,1743292800.0),(1048604,1743292800.0),(1048605,1743292800.0),(1048606,1743292800.0),(1048607,1743292800.0),(1048608,1743292800.0),(1048609,1743292800.0),(1048610,1743292800.0),(1048611,1743292800.0),(1048612,1743292800.0),(1048613,1743292800.0),(1048614,1743292800.0),(1048615,1743292800.0),(1048616,1743292800.0),(1048617,1743292800.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query INSERT INTO ttl_move_test VALUES (2097152,1743552000.0),(2097153,1743552000.0),(2097154,1743552000.0),(2097155,1743552000.0),(2097156,1743552000.0),(2097157,1743552000.0),(2097158,1743552000.0),(2097159,1743552000.0),(2097160,1743552000.0),(2097161,1743552000.0),(2097162,1743552000.0),(2097163,1743552000.0),(2097164,1743552000.0),(2097165,1743552000.0),(2097166,1743552000.0),(2097167,1743552000.0),(2097168,1743552000.0),(2097169,1743552000.0),(2097170,1743552000.0),(2097171,1743552000.0),(2097172,1743552000.0),(2097173,1743552000.0),(2097174,1743552000.0),(2097175,1743552000.0),(2097176,1743552000.0),(2097177,1743552000.0),(2097178,1743552000.0),(2097179,1743552000.0),(2097180,1743552000.0),(2097181,1743552000.0),(2097182,1743552000.0),(2097183,1743552000.0),(2097184,1743552000.0),(2097185,1743552000.0),(2097186,1743552000.0),(2097187,1743552000.0),(2097188,1743552000.0),(2097189,1743552000.0),(2097190,1743552000.0),(2097191,1743552000.0),(2097192,1743552000.0),(2097193,1743552000.0),( on node1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query OPTIMIZE TABLE ttl_move_test FINAL on node1 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SYSTEM SYNC REPLICA ttl_move_test on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node1 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT count() FROM ttl_move_test FORMAT Values on node2 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node1 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Executing query DROP TABLE IF EXISTS ttl_move_test SYNC on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml stop --timeout 20] [gw0] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3] Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/.env --project-name roottests3zerocopyreplication-gw0 --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --file /ClickHouse/tests/integration/test_s3_zero_copy_replication/_instances-0-gw0/node2/docker-compose.yml down --volumes] Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-node2-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-node1-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-resolver-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-zoo3-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-zoo2-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-zoo1-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-minio1-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Stopping Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Stopped Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Removing Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-proxy1-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Removed Stderr: Container roottests3zerocopyreplication-gw0-proxy2-1 Removed Stderr: Volume roottests3zerocopyreplication-gw0_data1-1 Removing Stderr: Volume roottests3zerocopyreplication-gw0_data1-1 Removing Stderr: Network roottests3zerocopyreplication-gw0_default Removing Stderr: Network roottests3zerocopyreplication-gw0_default Removing Stderr: Volume roottests3zerocopyreplication-gw0_data1-1 Removed Stderr: Volume roottests3zerocopyreplication-gw0_data1-1 Removed Stderr: Network roottests3zerocopyreplication-gw0_default Removed Stderr: Network roottests3zerocopyreplication-gw0_default Removed Cleanup called Cleanup called Docker networks for project roottests3zerocopyreplication-gw0 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottests3zerocopyreplication-gw0 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottests3zerocopyreplication-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottests3zerocopyreplication-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottests3zerocopyreplication-gw0 are DRIVER VOLUME NAME Docker volumes for project roottests3zerocopyreplication-gw0 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottests3zerocopyreplication-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottests3zerocopyreplication-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottests3zerocopyreplication-gw0 No running containers for project: roottests3zerocopyreplication-gw0 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:1 Stdout:1 Volumes pruned: 1 Volumes pruned: 1 Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Executing query SELECT errors_count, estimated_recovery_time FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] [gw6] PASSED test_system_clusters_actual_information/test.py::test Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] Command:[docker compose --env-file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/.env --project-name roottestsystemclustersactualinformation-gw6 --file /ClickHouse/tests/integration/test_system_clusters_actual_information/_instances-0-gw6/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Removing Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Removing Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Removed Stderr: Container roottestsystemclustersactualinformation-gw6-node-1 Removed Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Stopping Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Removing Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Removing Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Removing Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Removing Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Stopped Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Removing Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Removing Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Removed Stderr: Container roottestsystemclustersactualinformation-gw6-zoo2-1 Removed Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Removed Stderr: Container roottestsystemclustersactualinformation-gw6-zoo3-1 Removed Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Removed Stderr: Container roottestsystemclustersactualinformation-gw6-zoo1-1 Removed Stderr: Network roottestsystemclustersactualinformation-gw6_default Removing Stderr: Network roottestsystemclustersactualinformation-gw6_default Removing Stderr: Network roottestsystemclustersactualinformation-gw6_default Removed Stderr: Network roottestsystemclustersactualinformation-gw6_default Removed Cleanup called Cleanup called Docker networks for project roottestsystemclustersactualinformation-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestsystemclustersactualinformation-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestsystemclustersactualinformation-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestsystemclustersactualinformation-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestsystemclustersactualinformation-gw6 are DRIVER VOLUME NAME Docker volumes for project roottestsystemclustersactualinformation-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestsystemclustersactualinformation-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestsystemclustersactualinformation-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestsystemclustersactualinformation-gw6 No running containers for project: roottestsystemclustersactualinformation-gw6 Trying to prune unused networks... Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Stdout:Total reclaimed space: 0B Images pruned Images pruned Trying to prune unused volumes... Trying to prune unused volumes... Command:[docker volume ls | wc -l] Command:[docker volume ls | wc -l] Stdout:1 Stdout:1 Volumes pruned: 1 Volumes pruned: 1 =============================== warnings summary =============================== test_table_function_mongodb_legacy/test.py::test_auth_source[False] /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py:226: DeprecationWarning: add_user is deprecated and will be removed in PyMongo 4.0. Use db.command with createUser or updateUser instead admin_db.add_user( test_table_function_mongodb_legacy/test.py::test_complex_data_type[False] /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py:88: DeprecationWarning: add_user is deprecated and will be removed in PyMongo 4.0. Use db.command with createUser or updateUser instead db.add_user("root", "clickhouse") test_table_function_mongodb_legacy/test.py::test_complex_data_type[False] test_table_function_mongodb_legacy/test.py::test_secure_connection[True] test_table_function_mongodb_legacy/test.py::test_simple_select[False] /usr/local/lib/python3.10/dist-packages/pymongo/database.py:1247: DeprecationWarning: Creating a user with the read_only option or without roles is deprecated in MongoDB >= 2.6 warnings.warn("Creating a user with the read_only option " test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False] /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py:123: DeprecationWarning: add_user is deprecated and will be removed in PyMongo 4.0. Use db.command with createUser or updateUser instead db.add_user("root", "clickhouse") test_table_function_mongodb_legacy/test.py::test_missing_columns[False] /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py:262: DeprecationWarning: add_user is deprecated and will be removed in PyMongo 4.0. Use db.command with createUser or updateUser instead db.add_user("root", "clickhouse") test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False] /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py:185: DeprecationWarning: add_user is deprecated and will be removed in PyMongo 4.0. Use db.command with createUser or updateUser instead db.add_user("root", "clickhouse") test_throttling/test.py::test_backup_throttling[no_local_throttling] /usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py:1100: InsecureRequestWarning: Unverified HTTPS request is being made to host '172.16.7.4'. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/latest/advanced-usage.html#tls-warnings warnings.warn( test_s3_with_https/test.py::test_s3_with_https[s3_secure] /usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py:1100: InsecureRequestWarning: Unverified HTTPS request is being made to host '172.16.9.4'. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/latest/advanced-usage.html#tls-warnings warnings.warn( test_table_function_mongodb_legacy/test.py::test_secure_connection[True] /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py:144: DeprecationWarning: add_user is deprecated and will be removed in PyMongo 4.0. Use db.command with createUser or updateUser instead db.add_user("root", "clickhouse") test_table_function_mongodb_legacy/test.py::test_simple_select[False] /ClickHouse/tests/integration/test_table_function_mongodb_legacy/test.py:46: DeprecationWarning: add_user is deprecated and will be removed in PyMongo 4.0. Use db.command with createUser or updateUser instead db.add_user("root", "clickhouse") -- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html ============================== slowest durations =============================== 120.13s call test_system_clusters_actual_information/test.py::test 67.12s call test_store_cleanup/test.py::test_store_cleanup 65.29s setup test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection 37.07s call test_system_merges/test.py::test_mutation_simple[replicated] 36.39s call test_system_merges/test.py::test_mutation_simple[] 34.50s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3] 28.75s setup test_table_function_mongodb_legacy/test.py::test_simple_select[False] 28.32s setup test_table_function_mongodb_legacy/test.py::test_auth_source[False] 26.56s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 26.40s call test_system_logs_recreate/test.py::test_system_logs_recreate 26.18s teardown test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling] 25.64s setup test_table_function_mongodb_legacy/test.py::test_secure_connection[True] 24.38s setup test_storage_redis/test.py::test_create_table 22.94s setup test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference 21.69s teardown test_s3_table_functions/test.py::test_s3_table_functions_timeouts 21.63s teardown test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy] 21.62s teardown test_s3_cluster/test.py::test_wrong_cluster 21.56s teardown test_storage_url_with_proxy/test.py::test_s3_with_proxy_list 21.32s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3] 20.95s teardown test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards 20.53s setup test_system_clusters_actual_information/test.py::test 19.37s setup test_system_start_stop_listen/test.py::test_all_protocols 19.09s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3] 16.26s setup test_system_merges/test.py::test_merge_simple[] 16.09s call test_system_logs_recreate/test.py::test_drop_system_log 14.54s setup test_s3_with_https/test.py::test_s3_with_https[s3_secure] 14.48s setup test_system_queries/test.py::test_DROP_DNS_CACHE 14.39s setup test_s3_table_functions/test.py::test_s3_table_functions 14.31s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10] 14.16s setup test_storage_url/test.py::test_file_formats[CSV] 13.28s setup test_storage_url_with_proxy/test.py::test_s3_with_proxy_list 13.23s setup test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table] 13.11s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system 12.95s setup test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated 12.91s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter 12.68s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10] 12.50s call test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table] 12.20s call test_s3_table_functions/test.py::test_s3_table_functions_timeouts 12.19s call test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table] 12.15s setup test_storage_url_http_headers/test.py::test_storage_url_http_headers 11.92s setup test_ssh_keys_authentication/test.py::test_ecdsa 11.91s setup test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 11.88s setup test_structured_logging_json/test.py::test_structured_logging_json_format 11.83s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10] 10.64s call test_table_function_mongodb_legacy/test.py::test_simple_select[False] 9.97s call test_system_merges/test.py::test_merge_simple[] 9.84s call test_system_merges/test.py::test_merge_simple[replicated] 9.83s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3] 9.68s setup test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 9.40s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3] 9.37s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3] 9.13s teardown test_system_clusters_actual_information/test.py::test 7.96s setup test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side 6.94s setup test_throttling/test.py::test_backup_throttling[no_local_throttling] 6.75s teardown test_storage_url/test.py::test_url_cluster_with_named_collection 6.74s call test_system_start_stop_listen/test.py::test_all_protocols 6.39s setup test_system_logs_comment/test.py::test_system_logs_comment 6.11s setup test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance 6.00s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge 5.87s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 5.68s teardown test_system_merges/test.py::test_mutation_simple[replicated] 5.59s call test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter 5.48s teardown test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated 5.36s call test_system_start_stop_listen/test.py::test_except 5.32s teardown test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 5.21s teardown test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 5.07s teardown test_structured_logging_json/test.py::test_structured_logging_json_format 4.96s teardown test_system_queries/test.py::test_system_flush_logs 4.88s setup test_store_cleanup/test.py::test_store_cleanup 4.70s teardown test_system_start_stop_listen/test.py::test_except 4.61s teardown test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling] 4.58s teardown test_ssh_keys_authentication/test.py::test_wrong_key 4.58s teardown test_table_function_mongodb_legacy/test.py::test_simple_select[False] 4.57s teardown test_throttling/test.py::test_backup_throttling[no_local_throttling] 4.46s teardown test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers 4.31s teardown test_storage_redis/test.py::test_update 4.18s call test_system_start_stop_listen/test.py::test_default_protocols 4.09s teardown test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance 3.84s call test_system_logs_comment/test.py::test_system_logs_comment 3.40s call test_system_queries/test.py::test_system_flush_logs 2.95s teardown test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side 2.60s setup test_system_logs_recreate/test.py::test_drop_system_log 2.42s call test_system_queries/test.py::test_DROP_DNS_CACHE 2.25s call test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op] 2.18s call test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data 2.17s teardown test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table] 1.54s call test_table_function_mongodb_legacy/test.py::test_secure_connection[True] 1.46s call test_storage_azure_blob_storage/test_cluster.py::test_format_detection 1.36s call test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue 1.36s call test_system_start_stop_listen/test.py::test_custom_protocols 1.35s call test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op] 1.31s call test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op] 1.27s teardown test_system_logs_comment/test.py::test_system_logs_comment 1.19s call test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated 1.14s call test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op] 1.13s call test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY 1.11s teardown test_system_logs_recreate/test.py::test_system_logs_recreate 1.11s call test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op] 1.09s teardown test_store_cleanup/test.py::test_store_cleanup 1.06s call test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op] 0.89s call test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference 0.77s call test_storage_url/test.py::test_table_function_url_access_rights 0.65s call test_structured_logging_json/test.py::test_structured_logging_json_format 0.65s call test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table] 0.65s call test_storage_redis/test.py::test_create_table 0.62s call test_ssh_keys_authentication/test.py::test_key_with_passphrase 0.59s call test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling] 0.57s call test_ssh_keys_authentication/test.py::test_ecdsa 0.56s call test_storage_redis/test.py::test_simple_insert 0.55s call test_storage_url_http_headers/test.py::test_storage_url_http_headers 0.48s call test_storage_redis/test.py::test_delete 0.46s call test_storage_url/test.py::test_file_formats[TSV] 0.45s call test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection 0.41s call test_storage_url/test.py::test_file_formats[CSV] 0.41s call test_s3_with_https/test.py::test_s3_with_https[s3_secure] 0.41s call test_storage_url/test.py::test_file_formats[JSONEachRow] 0.40s call test_storage_url/test.py::test_partition_by 0.40s call test_storage_redis/test.py::test_update 0.38s call test_s3_table_functions/test.py::test_s3_table_functions 0.37s call test_storage_redis/test.py::test_select_int 0.36s call test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy] 0.35s call test_table_function_mongodb_legacy/test.py::test_complex_data_type[False] 0.35s call test_storage_url_http_headers/test.py::test_storage_url_redirected_headers 0.35s call test_storage_azure_blob_storage/test_cluster.py::test_select_all 0.34s call test_storage_url_with_proxy/test.py::test_s3_with_proxy_list 0.33s call test_s3_cluster/test.py::test_select_all 0.33s call test_storage_redis/test.py::test_truncate 0.33s call test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards 0.33s call test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers 0.31s call test_storage_url/test.py::test_file_formats[Parquet] 0.28s call test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling] 0.28s call test_s3_cluster/test.py::test_union_all 0.28s call test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards 0.28s call test_storage_redis/test.py::test_simple_select 0.27s call test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side 0.27s call test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase 0.26s call test_storage_azure_blob_storage/test_cluster.py::test_count 0.25s call test_storage_url/test.py::test_url_cluster 0.25s call test_storage_azure_blob_storage/test_cluster.py::test_union_all 0.24s call test_table_function_mongodb_legacy/test.py::test_auth_source[False] 0.23s call test_storage_url/test.py::test_url_cluster_with_named_collection 0.23s call test_throttling/test.py::test_backup_throttling[no_local_throttling] 0.22s call test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False] 0.20s call test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster 0.18s call test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS 0.14s call test_table_function_mongodb_legacy/test.py::test_no_credentials[False] 0.12s call test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False] 0.12s call test_s3_cluster/test.py::test_skip_unavailable_shards 0.11s call test_table_function_mongodb_legacy/test.py::test_missing_columns[False] 0.07s call test_ssh_keys_authentication/test.py::test_ed25519 0.07s call test_s3_cluster/test.py::test_unset_skip_unavailable_shards 0.07s call test_ssh_keys_authentication/test.py::test_wrong_key 0.07s call test_s3_cluster/test.py::test_wrong_cluster 0.07s call test_ssh_keys_authentication/test.py::test_rsa 0.02s call test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance 0.00s call test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance 0.00s setup test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling] 0.00s setup test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter 0.00s setup test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op] 0.00s setup test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op] 0.00s setup test_storage_url/test.py::test_url_cluster 0.00s setup test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10] 0.00s setup test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling] 0.00s setup test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table] 0.00s setup test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op] 0.00s setup test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op] 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3] 0.00s teardown test_system_merges/test.py::test_merge_simple[] 0.00s teardown test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table] 0.00s setup test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op] 0.00s setup test_system_merges/test.py::test_merge_simple[replicated] 0.00s setup test_storage_url/test.py::test_file_formats[JSONEachRow] 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 0.00s teardown test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10] 0.00s teardown test_table_function_mongodb_legacy/test.py::test_auth_source[False] 0.00s teardown test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op] 0.00s setup test_system_merges/test.py::test_mutation_simple[] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3] 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10] 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3] 0.00s setup test_storage_azure_blob_storage/test_cluster.py::test_union_all 0.00s setup test_storage_url/test.py::test_file_formats[TSV] 0.00s setup test_system_merges/test.py::test_mutation_simple[replicated] 0.00s teardown test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection 0.00s setup test_storage_redis/test.py::test_delete 0.00s teardown test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op] 0.00s setup test_storage_url/test.py::test_file_formats[Parquet] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation 0.00s setup test_storage_redis/test.py::test_simple_select 0.00s teardown test_system_logs_recreate/test.py::test_drop_system_log 0.00s setup test_system_start_stop_listen/test.py::test_default_protocols 0.00s setup test_table_function_mongodb_legacy/test.py::test_no_credentials[False] 0.00s setup test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance 0.00s setup test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers 0.00s teardown test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op] 0.00s setup test_storage_redis/test.py::test_simple_insert 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10] 0.00s setup test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS 0.00s teardown test_system_start_stop_listen/test.py::test_all_protocols 0.00s teardown test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance 0.00s setup test_storage_redis/test.py::test_truncate 0.00s setup test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards 0.00s setup test_ssh_keys_authentication/test.py::test_ed25519 0.00s setup test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False] 0.00s teardown test_system_queries/test.py::test_DROP_DNS_CACHE 0.00s teardown test_storage_url_http_headers/test.py::test_storage_url_http_headers 0.00s setup test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False] 0.00s teardown test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op] 0.00s teardown test_ssh_keys_authentication/test.py::test_ecdsa 0.00s setup test_s3_cluster/test.py::test_skip_unavailable_shards 0.00s teardown test_table_function_mongodb_legacy/test.py::test_secure_connection[True] 0.00s setup test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy] 0.00s teardown test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table] 0.00s setup test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase 0.00s setup test_storage_redis/test.py::test_select_int 0.00s teardown test_storage_url/test.py::test_file_formats[CSV] 0.00s setup test_storage_redis/test.py::test_update 0.00s setup test_ssh_keys_authentication/test.py::test_wrong_key 0.00s setup test_storage_azure_blob_storage/test_cluster.py::test_count 0.00s setup test_system_logs_recreate/test.py::test_system_logs_recreate 0.00s setup test_storage_url_http_headers/test.py::test_storage_url_redirected_headers 0.00s setup test_storage_azure_blob_storage/test_cluster.py::test_format_detection 0.00s teardown test_system_merges/test.py::test_merge_simple[replicated] 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3] 0.00s setup test_table_function_mongodb_legacy/test.py::test_complex_data_type[False] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 0.00s teardown test_storage_redis/test.py::test_create_table 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10] 0.00s teardown test_system_start_stop_listen/test.py::test_custom_protocols 0.00s setup test_storage_url/test.py::test_partition_by 0.00s teardown test_storage_url/test.py::test_partition_by 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge 0.00s setup test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system 0.00s teardown test_ssh_keys_authentication/test.py::test_rsa 0.00s teardown test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op] 0.00s teardown test_s3_with_https/test.py::test_s3_with_https[s3_secure] 0.00s setup test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter 0.00s teardown test_storage_url/test.py::test_file_formats[Parquet] 0.00s setup test_s3_cluster/test.py::test_select_all 0.00s setup test_system_start_stop_listen/test.py::test_custom_protocols 0.00s setup test_s3_table_functions/test.py::test_s3_table_functions_timeouts 0.00s teardown test_table_function_mongodb_legacy/test.py::test_complex_data_type[False] 0.00s teardown test_storage_redis/test.py::test_delete 0.00s teardown test_s3_table_functions/test.py::test_s3_table_functions 0.00s setup test_storage_azure_blob_storage/test_cluster.py::test_select_all 0.00s setup test_table_function_mongodb_legacy/test.py::test_missing_columns[False] 0.00s setup test_storage_url/test.py::test_table_function_url_access_rights 0.00s setup test_ssh_keys_authentication/test.py::test_rsa 0.00s setup test_s3_cluster/test.py::test_union_all 0.00s setup test_ssh_keys_authentication/test.py::test_key_with_passphrase 0.00s teardown test_storage_url_http_headers/test.py::test_storage_url_redirected_headers 0.00s setup test_s3_cluster/test.py::test_wrong_cluster 0.00s setup test_system_start_stop_listen/test.py::test_except 0.00s teardown test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op] 0.00s teardown test_storage_redis/test.py::test_simple_insert 0.00s teardown test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards 0.00s setup test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY 0.00s setup test_s3_cluster/test.py::test_unset_skip_unavailable_shards 0.00s teardown test_table_function_mongodb_legacy/test.py::test_missing_columns[False] 0.00s setup test_system_queries/test.py::test_system_flush_logs 0.00s teardown test_storage_url/test.py::test_file_formats[JSONEachRow] 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system 0.00s teardown test_storage_url/test.py::test_file_formats[TSV] 0.00s teardown test_storage_azure_blob_storage/test_cluster.py::test_select_all 0.00s setup test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter 0.00s teardown test_storage_url/test.py::test_table_function_url_access_rights 0.00s teardown test_system_merges/test.py::test_mutation_simple[] 0.00s setup test_storage_url/test.py::test_url_cluster_with_named_collection 0.00s teardown test_storage_redis/test.py::test_simple_select 0.00s teardown test_storage_redis/test.py::test_select_int 0.00s teardown test_storage_redis/test.py::test_truncate 0.00s teardown test_s3_cluster/test.py::test_select_all 0.00s teardown test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS 0.00s teardown test_storage_azure_blob_storage/test_cluster.py::test_count 0.00s teardown test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY 0.00s teardown test_table_function_mongodb_legacy/test.py::test_no_credentials[False] 0.00s teardown test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge 0.00s teardown test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False] 0.00s teardown test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster 0.00s teardown test_ssh_keys_authentication/test.py::test_ed25519 0.00s teardown test_ssh_keys_authentication/test.py::test_key_with_passphrase 0.00s teardown test_storage_url/test.py::test_url_cluster 0.00s teardown test_system_start_stop_listen/test.py::test_default_protocols 0.00s teardown test_storage_azure_blob_storage/test_cluster.py::test_format_detection 0.00s teardown test_s3_cluster/test.py::test_union_all 0.00s teardown test_s3_cluster/test.py::test_unset_skip_unavailable_shards 0.00s teardown test_storage_azure_blob_storage/test_cluster.py::test_union_all 0.00s teardown test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False] 0.00s teardown test_s3_cluster/test.py::test_skip_unavailable_shards 0.00s teardown test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase =========================== short test summary info ============================ PASSED test_ssh_keys_authentication/test.py::test_ecdsa PASSED test_ssh_keys_authentication/test.py::test_ed25519 PASSED test_ssh_keys_authentication/test.py::test_key_with_passphrase PASSED test_ssh_keys_authentication/test.py::test_key_with_wrong_passphrase PASSED test_ssh_keys_authentication/test.py::test_rsa PASSED test_ssh_keys_authentication/test.py::test_wrong_key PASSED test_storage_url/test.py::test_file_formats[CSV] PASSED test_storage_url/test.py::test_file_formats[JSONEachRow] PASSED test_storage_url/test.py::test_file_formats[Parquet] PASSED test_storage_url/test.py::test_file_formats[TSV] PASSED test_storage_url/test.py::test_partition_by PASSED test_system_queries/test.py::test_DROP_DNS_CACHE PASSED test_storage_url/test.py::test_table_function_url_access_rights PASSED test_system_queries/test.py::test_RELOAD_CONFIG_AND_MACROS PASSED test_storage_url/test.py::test_url_cluster PASSED test_storage_url/test.py::test_url_cluster_with_named_collection PASSED test_system_queries/test.py::test_SYSTEM_RELOAD_DICTIONARY PASSED test_system_queries/test.py::test_system_flush_logs PASSED test_s3_cluster/test.py::test_parallel_distributed_insert_select_with_schema_inference PASSED test_s3_cluster/test.py::test_select_all PASSED test_s3_cluster/test.py::test_skip_unavailable_shards PASSED test_s3_cluster/test.py::test_union_all PASSED test_s3_cluster/test.py::test_unset_skip_unavailable_shards PASSED test_s3_cluster/test.py::test_wrong_cluster PASSED test_storage_redis/test.py::test_create_table PASSED test_storage_redis/test.py::test_delete PASSED test_storage_redis/test.py::test_select_int PASSED test_system_start_stop_listen/test.py::test_all_protocols PASSED test_system_merges/test.py::test_merge_simple[] PASSED test_storage_redis/test.py::test_simple_insert PASSED test_storage_redis/test.py::test_simple_select PASSED test_storage_redis/test.py::test_truncate PASSED test_storage_redis/test.py::test_update PASSED test_system_start_stop_listen/test.py::test_custom_protocols PASSED test_table_function_mongodb_legacy/test.py::test_auth_source[False] PASSED test_table_function_mongodb_legacy/test.py::test_complex_data_type[False] PASSED test_table_function_mongodb_legacy/test.py::test_incorrect_data_type[False] PASSED test_table_function_mongodb_legacy/test.py::test_missing_columns[False] PASSED test_table_function_mongodb_legacy/test.py::test_no_credentials[False] PASSED test_table_function_mongodb_legacy/test.py::test_predefined_connection_configuration[False] PASSED test_storage_url_http_headers/test.py::test_storage_url_http_headers PASSED test_storage_url_http_headers/test.py::test_storage_url_redirected_headers PASSED test_storage_url_http_headers/test.py::test_with_override_content_type_url_http_headers PASSED test_system_start_stop_listen/test.py::test_default_protocols PASSED test_throttling/test.py::test_backup_throttling[no_local_throttling] PASSED test_system_merges/test.py::test_merge_simple[replicated] PASSED test_system_start_stop_listen/test.py::test_except PASSED test_throttling/test.py::test_backup_throttling[no_local_to_remote_throttling] PASSED test_throttling/test.py::test_backup_throttling[no_remote_to_local_throttling] PASSED test_server_keep_alive/test.py::test_max_keep_alive_requests_on_user_side PASSED test_s3_with_https/test.py::test_s3_with_https[s3_secure] PASSED test_s3_with_https/test.py::test_s3_with_https[s3_secure_with_proxy] PASSED test_system_detached_tables/test.py::test_system_detached_tables[Atomic db with MergeTree table] PASSED test_system_detached_tables/test.py::test_system_detached_tables[Lazy db with Log table] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_replication[s3] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_after_op] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[None-zero_copy_lock_zk_fail_before_op] PASSED test_table_function_mongodb_legacy/test.py::test_secure_connection[True] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_after_op] PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_after_op-zero_copy_lock_zk_fail_before_op] PASSED test_system_logs_recreate/test.py::test_drop_system_log PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_after_op] PASSED test_storage_url_with_proxy/test.py::test_s3_with_proxy_list PASSED test_s3_table_functions/test.py::test_s3_table_functions PASSED test_s3_zero_copy_replication/test.py::test_move_shared_zero_copy_lock_fail[zero_copy_unlock_zk_fail_before_op-zero_copy_lock_zk_fail_before_op] PASSED test_system_detached_tables/test.py::test_system_detached_tables[Replicated db with ReplicatedMergeTree table] PASSED test_storage_azure_blob_storage/test_cluster.py::test_cluster_with_named_collection PASSED test_storage_azure_blob_storage/test_cluster.py::test_count PASSED test_storage_azure_blob_storage/test_cluster.py::test_format_detection PASSED test_storage_azure_blob_storage/test_cluster.py::test_partition_parallel_reading_with_cluster PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_concurrent_merge PASSED test_storage_azure_blob_storage/test_cluster.py::test_select_all PASSED test_storage_azure_blob_storage/test_cluster.py::test_skip_unavailable_shards PASSED test_storage_azure_blob_storage/test_cluster.py::test_union_all PASSED test_storage_azure_blob_storage/test_cluster.py::test_unset_skip_unavailable_shards PASSED test_system_merges/test.py::test_mutation_simple[] PASSED test_s3_table_functions/test.py::test_s3_table_functions_timeouts PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_alter PASSED test_system_ddl_worker_queue/test.py::test_distributed_ddl_queue PASSED test_system_logs_recreate/test.py::test_system_logs_recreate PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_drop_detached_system PASSED test_structured_logging_json/test.py::test_structured_logging_json_format PASSED test_table_function_mongodb_legacy/test.py::test_simple_select[False] PASSED test_system_logs_comment/test.py::test_system_logs_comment PASSED test_system_logs_hostname/test_replicated.py::test_truncate_database_replicated PASSED test_tcp_handler_http_responses/test_case.py::test_request_to_http_full_instance PASSED test_tcp_handler_http_responses/test_case.py::test_request_to_http_less_instance PASSED test_temporary_data_in_cache/test.py::test_cache_evicted_by_temporary_data PASSED test_system_merges/test.py::test_mutation_simple[replicated] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_keeps_data_after_mutation PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_alter PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_unfreeze_system PASSED test_store_cleanup/test.py::test_store_cleanup PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[False-10] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_delete[True-3] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-False-10] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered-True-3] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-False-10] PASSED test_s3_zero_copy_replication/test.py::test_s3_zero_copy_with_ttl_move[tiered_copy-True-3] PASSED test_system_clusters_actual_information/test.py::test ================= 100 passed, 12 warnings in 267.22s (0:04:27) =================